This new API helps customers to better estimate and measure Job usage for billing or monitoring.

This API retrieves the DVU count for executed jobs. Results can be grouped either by DataSource or JobId (default), depending on the parameter group_by.

To see the documentation for the the Jobs DVU Count API, you can access it here: https://docs.datastreamer.io/docs/jobs-dvu-count-api#/

With the release of Datastreamer's MCP server, you can now integrate your AI features closer to your data pipelines!

If you have not read the "CTO Brief" on Datastreamer's strategy of being the agent interface for social data, you can access it here: https://datastreamer.io/agent-interface-for-social-data-the-cto-edition/

Inside this update, you can now access the Datastreamer MCP server and the first of our AI tools "Create Job" which allows you to create data collection jobs with natural language. This gives your new AI features, the ability to easily access the data they need.

Read more about the Create Job tool here: https://docs.datastreamer.io/docs/job-creation-agent#/

Get started connecting your MCP client here: https://docs.datastreamer.io/docs/mcp-server-setup-guide#/

When you create a Periodic Job, the default behaviour is to collect any new content matching your query since the last run time. Some customers, working to identify changes in the content's metrics, have requested greater flexibility in coverage time.

With the "Query Start Time Adjustment" option in all Periodic Jobs, you can now set an adjusted search start time. You can now set the number of seconds to adjust the query start time backwards in time. For example: search hourly, for the previous two hours. This is best used with updatable storage (like Searchable Storage) to ensure updated fields are registered.

  • Greater support for custom data sources: https://docs.datastreamer.io/update/docs/private-data-sources#/
  • Improvements to Pipeline validation, improving the error messages when running misconfigured pipelines.
  • Removed empty and unnecessary fields from Pipeline export files.
  • Fixed issues with query builder when using select Socialgist sources.
  • Improved chunking when using GenAI Translation component.

Within Datastreamer, you can now integrate any Apify actor into your pipelines!

About Apify: Apify provides a marketplace of scrapers designed for large-scale web scraping, data extraction, and browser automation. This marketplace of community-made scrapers, let's you run 5,000+ different scrapers known as "Actors".

The new Apify Actor Integration component allows you to use any of the Actors from Apify within your pipeline.

To get started, check out this Setup Guide: Apify Integration Setup.

Not sure what to start with? Registry offers a selection of the 5,000+ Apify Actors: