DataflowBundle is a bundle for Symfony 3.4+ providing an easy way to create import / export dataflow.
Dataflow | Symfony | Support |
---|---|---|
5.x | 7.x | yes |
4.x | 3.4 | 4.x | 5.x | 6.x | yes |
3.x | 3.4 | 4.x | 5.x | no |
2.x | 3.4 | 4.x | no |
1.x | 3.4 | 4.x | no |
Dataflow uses a linear generic workflow in three parts:
- one reader
- any number of steps that can be synchronous or asynchronous
- one or more writers
The reader can read data from anywhere and return data row by row. Each step processes the current row data. The steps are executed in the order in which they are added. And, one or more writers save the row anywhere you want.
As the following schema shows, you can define more than one dataflow:
- Define and configure a Dataflow
- Run the Job scheduled
- Run one Dataflow from the command line
- Define the schedule for a Dataflow from the command line
- Enable/Disable a scheduled Dataflow from the command line
- Display the list of scheduled Dataflow from the command line
- Display the result for the last Job for a Dataflow from the command line
- Work with multiple Doctrine DBAL connections
Security notice: Symfony 4.x is not supported before 4.1.12, see https://github.com/advisories/GHSA-pgwj-prpq-jpc2
To install this bundle, run this command :
$ composer require code-rhapsodie/dataflow-bundle
You can use the generic readers, writers and steps from PortPHP.
For the writers, you must use the adapter CodeRhapsodie\DataflowBundle\DataflowType\Writer\PortWriterAdapter
like
this:
<?php
// ...
$streamWriter = new \Port\Writer\StreamMergeWriter();
$builder->addWriter(new \CodeRhapsodie\DataflowBundle\DataflowType\Writer\PortWriterAdapter($streamWriter));
// ...
Add CodeRhapsodie\DataflowBundle\CodeRhapsodieDataflowBundle::class => ['all' => true],
in the config/bundles.php
file.
Like this:
<?php
return [
// ...
CodeRhapsodie\DataflowBundle\CodeRhapsodieDataflowBundle::class => ['all' => true],
// ...
];
This bundle uses Doctrine DBAL to store Dataflow schedule into the database table (cr_dataflow_scheduled
)
and jobs (cr_dataflow_job
).
If you use Doctrine Migration Bundle or Phinx or Kaliop Migration Bundle or whatever, you can add a new migration with the generated SQL query from this command:
$ bin/console code-rhapsodie:dataflow:dump-schema
If you have already the tables, you can add a new migration with the generated update SQL query from this command:
$ bin/console code-rhapsodie:dataflow:dump-schema --update
By default, the Doctrine DBAL connection used is default
. You can configure the default connection.
Add this configuration into your Symfony configuration:
code_rhapsodie_dataflow:
dbal_default_connection: test #Name of the default connection used by Dataflow bundle
By default, the logger
service will be used to log all exceptions and custom messages.
If you want to use another logger, like a specific Monolog handler, Add this configuration:
code_rhapsodie_dataflow:
default_logger: monolog.logger.custom #Service ID of the logger you want Dataflow to use
Dataflow can delegate the execution of its jobs to the Symfony messenger component, if available. This allows jobs to be executed concurrently by workers instead of sequentially.
To enable messenger mode:
code_rhapsodie_dataflow:
messenger_mode:
enabled: true
# bus: 'messenger.default_bus' #Service ID of the bus you want Dataflow to use, if not the default one
You also need to route Dataflow messages to the proper transport:
# config/packages/messenger.yaml
framework:
messenger:
transports:
async: '%env(MESSENGER_TRANSPORT_DSN)%'
routing:
CodeRhapsodie\DataflowBundle\MessengerMode\JobMessage: async
This bundle uses a fixed and simple workflow structure in order to let you focus on the data processing logic part of your dataflow.
A dataflow type defines the different parts of your dataflow. A dataflow is made of:
- exactly one Reader
- any number of Steps
- one or more Writers
Dataflow types can be configured with options.
A dataflow type must implement CodeRhapsodie\DataflowBundle\DataflowType\DataflowTypeInterface
.
To help with creating your dataflow types, an abstract
class CodeRhapsodie\DataflowBundle\DataflowType\AbstractDataflowType
is provided, allowing you to define your dataflow through a handy
builder CodeRhapsodie\DataflowBundle\DataflowType\DataflowBuilder
.
This is an example to define one class DataflowType:
<?php
namespace CodeRhapsodie\DataflowExemple\DataflowType;
use CodeRhapsodie\DataflowBundle\DataflowType\AbstractDataflowType;
use CodeRhapsodie\DataflowBundle\DataflowType\DataflowBuilder;
use CodeRhapsodie\DataflowExemple\Reader\FileReader;
use CodeRhapsodie\DataflowExemple\Writer\FileWriter;
class MyFirstDataflowType extends AbstractDataflowType
{
private $myReader;
private $myWriter;
public function __construct(FileReader $myReader, FileWriter $myWriter)
{
$this->myReader = $myReader;
$this->myWriter = $myWriter;
}
protected function buildDataflow(DataflowBuilder $builder, array $options): void
{
$this->myWriter->setDestinationFilePath($options['to-file']);
$builder
->setReader($this->myReader->read($options['from-file']))
->addStep(function ($data) use ($options) {
// TODO : Write your code here...
return $data;
})
->addWriter($this->myWriter)
;
}
protected function configureOptions(OptionsResolver $optionsResolver): void
{
$optionsResolver->setDefaults(['to-file' => '/tmp/dataflow.csv', 'from-file' => null]);
$optionsResolver->setRequired('from-file');
}
public function getLabel(): string
{
return 'My First Dataflow';
}
public function getAliases(): iterable
{
return ['mfd'];
}
}
Dataflow types must be tagged with coderhapsodie.dataflow.type
.
If you're using Symfony auto-configuration for your services, this tag will be automatically added to all services
implementing DataflowTypeInterface
.
Otherwise, manually add the tag coderhapsodie.dataflow.type
in your dataflow type service configuration:
```yaml
CodeRhapsodie\DataflowExemple\DataflowType\MyFirstDataflowType:
tags:
- { name: coderhapsodie.dataflow.type }
The AbstractDataflowType
can help you define options for your Dataflow type.
Add this method in your DataflowType class:
<?php
// ...
use Symfony\Component\OptionsResolver\OptionsResolver;
class MyFirstDataflowType extends AbstractDataflowType
{
// ...
protected function configureOptions(OptionsResolver $optionsResolver): void
{
$optionsResolver->setDefaults(['to-file' => '/tmp/dataflow.csv', 'from-file' => null]);
$optionsResolver->setRequired('from-file');
}
}
With this configuration, the option fileName
is required. For an advanced usage of the option resolver, read
the Symfony documentation.
For asynchronous management, AbstractDataflowType
come with two default options :
- loopInterval : default to 0. Update this interval if you wish customise the
tick
loop duration. - emitInterval : default to 0. Update this interval to have a control when reader must emit new data in the flow pipeline.
All exceptions will be caught and written in the logger.
If you want to add custom messages in the log, you can inject the logger in your readers / steps / writers.
If your DataflowType class extends AbstractDataflowType
, the logger is accessible as $this->logger
.
<?php
// ...
use Symfony\Component\OptionsResolver\OptionsResolver;
class MyDataflowType extends AbstractDataflowType
{
// ...
protected function buildDataflow(DataflowBuilder $builder, array $options): void
{
$this->myWriter->setLogger($this->logger);
}
}
When using the code-rhapsodie:dataflow:run-pending
command, this logger will also be used to save the log in the
corresponding job in the database.
Execute this command to check if your DataflowType is correctly registered:
$ bin/console debug:container --tag coderhapsodie.dataflow.type
The result is like this:
Symfony Container Public and Private Services Tagged with "coderhapsodie.dataflow.type" Tag
===========================================================================================
---------------------------------------------------------------- ----------------------------------------------------------------
Service ID Class name
---------------------------------------------------------------- ----------------------------------------------------------------
CodeRhapsodie\DataflowExemple\DataflowType\MyFirstDataflowType CodeRhapsodie\DataflowExemple\DataflowType\MyFirstDataflowType
---------------------------------------------------------------- ----------------------------------------------------------------
Readers provide the dataflow with elements to import / export. Usually, elements are read from an external resource ( file, database, webservice, etc).
A Reader can be any iterable
.
The only constraint on the returned elements typing is that they cannot be false
.
The reader can be a generator like this example :
<?php
namespace CodeRhapsodie\DataflowExemple\Reader;
class FileReader
{
public function read(string $filename): iterable
{
if (!$filename) {
throw new \Exception("The file name is not defined. Define it with 'setFilename' method");
}
if (!$fh = fopen($filename, 'r')) {
throw new \Exception("Unable to open file '".$filename."' for read.");
}
while (false !== ($read = fgets($fh))) {
yield explode('|', trim($read));
}
}
}
You can set up this reader as follows:
$builder->setReader(($this->myReader)())
Steps are operations performed on the elements before they are handled by the Writers. Usually, steps are either:
- converters, that alter the element
- filters, that conditionally prevent further operations on the element
- generators, that can include asynchronous operations
A Step can be any callable, taking the element as its argument, and returning either:
- the element, possibly altered
false
, if no further operations should be performed on this element
A few examples:
<?php
//[...]
$builder->addStep(function ($item) {
// Titles are changed to all caps before export
$item['title'] = strtoupper($item['title']);
return $item;
});
// asynchronous step with 2 scale factor
$builder->addStep(function ($item): \Generator {
yield new \Amp\Delayed(1000); // asynchronous processing for 1 second long
// Titles are changed to all caps before export
$item['title'] = strtolower($item['title']);
return $item;
}, 2);
$builder->addStep(function ($item) {
// Private items are not exported
if ($item['private']) {
return false;
}
return $item;
});
//[...]
Note : you can ensure writing order for asynchronous operations if all steps are scaled at 1 factor.
Writers perform the actual import / export operations.
A Writer must implement CodeRhapsodie\DataflowBundle\DataflowType\Writer\WriterInterface
.
As this interface is not compatible with Port\Writer
, the
adapter CodeRhapsodie\DataflowBundle\DataflowType\Writer\PortWriterAdapter
is provided.
This example show how to use the predefined PhpPort Writer :
$builder->addWriter(new PortWriterAdapter(new \Port\FileWriter()));
Or your own Writer:
<?php
namespace CodeRhapsodie\DataflowExemple\Writer;
use CodeRhapsodie\DataFlowBundle\DataflowType\Writer\WriterInterface;
class FileWriter implements WriterInterface
{
private $fh;
/** @var string */
private $path;
public function setDestinationFilePath(string $path) {
$this->path = $path;
}
public function prepare()
{
if (null === $this->path) {
throw new \Exception('Define the destination file name before use');
}
if (!$this->fh = fopen($this->path, 'w')) {
throw new \Exception('Unable to open in write mode the output file.');
}
}
public function write($item)
{
fputcsv($this->fh, $item);
}
public function finish()
{
fclose($this->fh);
}
}
If you want to write multiple items from a single item read, you can use the generic CollectionWriter
. This writer
will iterate over any iterable
it receives, and pass each item from that collection to your own writer that handles
single items.
$builder->addWriter(new CollectionWriter($mySingleItemWriter));
If you want to call different writers depending on what item is read, you can use the generic DelegatorWriter
.
As an example, let's suppose our items are arrays with the first entry being either product
or order
. We want to use
a different writer based on that value.
First, create your writers implementing DelegateWriterInterface
(this interface extends WriterInterface
so your
writers can still be used without the DelegatorWriter
).
<?php
namespace CodeRhapsodie\DataflowExemple\Writer;
use CodeRhapsodie\DataFlowBundle\DataflowType\Writer\WriterInterface;
class ProductWriter implements DelegateWriterInterface
{
public function supports($item): bool
{
return 'product' === reset($item);
}
public function prepare()
{
}
public function write($item)
{
// Process your product
}
public function finish()
{
}
}
<?php
namespace CodeRhapsodie\DataflowExemple\Writer;
use CodeRhapsodie\DataFlowBundle\DataflowType\Writer\WriterInterface;
class OrderWriter implements DelegateWriterInterface
{
public function supports($item): bool
{
return 'order' === reset($item);
}
public function prepare()
{
}
public function write($item)
{
// Process your order
}
public function finish()
{
}
}
Then, configure your DelegatorWriter
and add it to your dataflow type.
protected function buildDataflow(DataflowBuilder $builder, array $options): void
{
// Snip add reader and steps
$delegatorWriter = new DelegatorWriter();
$delegatorWriter->addDelegate(new ProductWriter());
$delegatorWriter->addDelegate(new OrderWriter());
$builder->addWriter($delegatorWriter);
}
During execution, the DelegatorWriter
will simply pass each item received to its first delegate (in the order those
were added) that supports it. If no delegate supports an item, an exception will be thrown.
All pending dataflow job processes are stored in a queue into the database.
Add this command into your crontab for execute all queued jobs:
$ SYMFONY_ENV=prod php bin/console code-rhapsodie:dataflow:run-pending
Several commands are provided to manage schedules and run jobs.
code-rhapsodie:dataflow:run-pending
Executes job in the queue according to their schedule.
When messenger mode is enabled, jobs will still be created according to their schedule, but execution will be handled by the messenger component instead.
code-rhapsodie:dataflow:schedule:list
Display the list of dataflows scheduled.
code-rhapsodie:dataflow:schedule:change-status
Enable or disable a scheduled dataflow
code-rhapsodie:dataflow:schedule:add
Add the schedule for a dataflow.
code-rhapsodie:dataflow:job:show
Display the last result of a job.
code-rhapsodie:dataflow:execute
Let you execute one dataflow job.
code-rhapsodie:dataflow:dump-schema
Generates schema create / update SQL queries
All commands have a --connection
option to define what Doctrine DBAL connection to use during execution.
Example:
This command uses the default
DBAL connection to generate all schema update queries.
$ bin/console code-rhapsodie:dataflow:dump-schema --update --connection=default
To execute all pending job for a specific connection use:
# Run for dataflow DBAL connection
$ bin/console code-rhapsodie:dataflow:run-pending --connection=dataflow
# Run for default DBAL connection
$ bin/console code-rhapsodie:dataflow:run-pending --connection=default
Please report issues and request features at https://github.com/code-rhapsodie/dataflow-bundle/issues.
Please note that only the last release of the 4.x and the 5.x versions of this bundle are actively supported.
Contributions are very welcome. Please see CONTRIBUTING.md for details. Thanks to everyone who has contributed already.
This package is licensed under the MIT license.