The Documents plugin provides functionality for generating, persisting, combining, and manipulating documents within the FlowX.AI system.
values.yaml
file:
SECURITY_OAUTH2_BASE_SERVER_URL
SECURITY_OAUTH2_CLIENT_CLIENT_ID
SECURITY_OAUTH2_REALM
FLOWX_HTML_TEMPLATES_ENABLED
environment variable to true.
SPRING_DATASOURCE_URL
: The URL for the Postgres/Oracle database.SPRING_DATASOURCE_USERNAME
: The username for the database connection.SPRING_DATASOURCE_PASSWORD
: The password for the database connection.SPRING_JPA_PROPERTIES_HIBERNATE_DEFAULT_SCHEMA
: Use this property to overwrite the name of the database schema if needed.SPRING_DATA_MONGODB_URI
environment variable to the MongoDB database URI.
SPRING_REDIS_HOST
: The host address of the Redis server.SPRING_REDIS_PASSWORD
: The password for the Redis server, if applicable.REDIS_TTL
: The time-to-live (TTL) value for Redis cache entries.FLOWX_CONVERT_DPI
: Sets the DPI (dots per inch) for PDF to JPEG conversion. Higher values result in higher resolution images. (Default value: 150
).SPRING_KAFKA_BOOTSTRAP_SERVERS
: The address of the Kafka server.SPRING_KAFKA_CONSUMER_GROUP_ID
: The group ID for Kafka consumers.KAFKA_CONSUMER_THREADS
: The number of Kafka consumer threads to use.KAFKA_AUTH_EXCEPTION_RETRY_INTERVAL
: The interval between retries after a AuthorizationException
is thrown by KafkaConsumer
.KAFKA_MESSAGE_MAX_BYTES
: The maximum size of a message that can be received by the broker from a producer.KAFKA_TOPIC_DOCUMENT_GENERATE_HTML_IN
: This Kafka topic is used for messages related to generating HTML documents (the topic that listens for the request from the engine)KAFKA_TOPIC_DOCUMENT_GENERATE_HTML_OUT
: This Kafka topic is used for messages related to generating HTML documents (the topic on which the engine will expect the reply)KAFKA_TOPIC_DOCUMENT_GENERATE_PDF_IN
: This Kafka topic is used for the input messages related to generating PDF documents (the topic that listens for the request from the engine)KAFKA_TOPIC_DOCUMENT_GENERATE_PDF_OUT
: This Kafka topic is used for the output messages related to generating PDF documents, it produces messages with the result of generating a PDF document (the topic on which the engine will expect the reply)KAFKA_TOPIC_FILE_PERSIST_IN
: This Kafka topic is used for the input messages related to persisting files, it receives messages indicating the request to persist a file (the topic that listens for the request from the engine)KAFKA_TOPIC_FILE_PERSIST_OUT
: This Kafka topic is used for the output messages related to persisting files, it produces messages with the result of persisting a file (the topic on which the engine will expect the reply)KAFKA_TOPIC_DOCUMENT_PERSIST_IN
: This Kafka topic is used for the input messages related to persisting documents, it receives messages indicating the request to persist a document (the topic that listens for the request from the engine)KAFKA_TOPIC_DOCUMENT_PERSIST_OUT
: This Kafka topic is used for the output messages related to persisting documents, it produces messages with the result of persisting a document (the topic that listens for the request from the engine)KAFKA_TOPIC_DOCUMENT_SPLIT_IN
: This Kafka topic is used for the input messages related to splitting documents, it receives messages indicating the request to split a document into multiple parts (the topic that listens for the request from the engine)KAFKA_TOPIC_DOCUMENT_SPLIT_OUT
: This Kafka topic is used for the output messages related to splitting documents, it produces messages with the result of splitting a document (the topic on which the engine will expect the reply)KAFKA_TOPIC_FILE_COMBINE_IN
: This Kafka topic is used for the input messages related to combining files, it receives messages indicating the request to combine multiple files into a single file (the topic that listens for the request from the engine)KAFKA_TOPIC_FILE_COMBINE_OUT
: This Kafka topic is used for the output messages related to combining files, it produces messages with the result of combining files (the topic on which the engine will expect the reply)KAFKA_TOPIC_DOCUMENT_GET_URLS_IN
: This Kafka topic is used for the input messages related to retrieving URLs for documents, it receives messages indicating the request to retrieve the URLs of documents (the topic that listens for the request from the engine)KAFKA_TOPIC_DOCUMENT_GET_URLS_OUT
: This Kafka topic is used for the output messages related to retrieving URLs for documents, it produces messages with the result of retrieving the URLs of documents (the topic on which the engine will expect the reply)KAFKA_TOPIC_FILE_DELETE_IN
: This Kafka topic is used for the input messages related to deleting files, it receives messages indicating the request to delete a file (the topic that listens for the request from the engine)KAFKA_TOPIC_FILE_DELETE_OUT
: This Kafka topic is used for the output messages related to deleting files, it produces messages with the result of deleting a file (the topic on which the engine will expect the reply)KAFKA_TOPIC_OCR_OUT
: This Kafka topic is used for the output messages related to optical character recognition (OCR), it produces messages with the OCR results (the topic on which the engine will expect the reply)KAFKA_TOPIC_OCR_IN
: This Kafka topic is used for the input messages related to optical character recognition (OCR), it receives messages indicating the request to perform OCR on a document (the topic that listens for the request from the engine)APPLICATION_FILE_STORAGE_PARTITION_STRATEGY
: Set the partition strategy for file storage. Use NONE
to save documents in minio/amazon-s3
as before, with a bucket for each process instance. Use PROCESS_DATE
to save documents in a single bucket with a subfolder structure, for example: bucket/2022/2022-07-04/process-id-xxxx/customer-id/file.pdf
.APPLICATION_FILE_STORAGE_DELETION_STRATEGY
(default value: delete): This will keep the current behaviour of deleting the temporary files.
Other possible values:
x-amz-bypass-governance-retention:true
, to enable deletion of governed files, in case the s3 configured user for document-plugin, will have the s3:BypassGovernanceRetention
permission.APPLICATION_FILE_STORAGE_S3_SERVER_URL
: The URL of the S3-compatible server.APPLICATION_FILE_STORAGE_S3_ACCESS_KEY
: The access key for the S3-compatible server.APPLICATION_FILE_STORAGE_S3_SECRET_KEY
: The secret key for the S3-compatible server.APPLICATION_FILE_STORAGE_S3_BUCKET_PREFIX
: The prefix to use for S3 bucket names.APPLICATION_FILE_STORAGE_S3_TEMP_BUCKET
: Upon file upload, the initial destination is a sandbox, from which it is subsequently transferred to the designated bucket.SPRING_SERVLET_MULTIPART_MAX_FILE_SIZE
and SPRING_SERVLET_MULTIPART_MAX_REQUEST_SIZE
variables.
FLOWX_HTML_TEMPLATES_PDF_FONT_PATHS
config to select the font used for generating documents based on PDF templates.
FLOWX_HTML_TEMPLATES_PDF_FONT_PATHS
config. By default, Calibri and DejaVuSans are available fonts.
After making these configurations, the fonts will be available for use within PDF templates.
LOGGING_LEVEL_ROOT
: Controls the log level for root Spring Boot microservice logs.LOGGING_LEVEL_APP
: Controls the log level for application-specific logs.LOGGING_LEVEL_MONGO_DRIVER
: Controls the log level for MongoDB driver logs.