Publishing Batches Of Events

Fiddler has a flexible ETL framework for retrieving and publishing batches of production data, either from local storage or from the cloud. This provides maximum flexibility in how you are required to store your data when publishing events to Fiddler.

The following data formats are currently supported:

  • pandas DataFrame objects (pd.DataFrame)

  • CSV files (.csv),

  • Parquet files (.pq)

  • Pickled pandas DataFrame objects (.pkl),

  • gzipped CSV files (.csv.gz),

The following data locations are supported:

  • In memory (for DataFrames)

  • Local disk

  • AWS S3

Once you have a batch of events stored somewhere, all you need to do to publish the batch to Fiddler is call the Fiddler client's fdl.Model.publish function.


After calling the function, please allow 3-5 minutes for events to populate the Monitor page.

Last updated

© 2024 Fiddler AI