[ 
https://issues.apache.org/jira/browse/CAMEL-8211?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15175514#comment-15175514
 ] 

Luca Burgazzoli commented on CAMEL-8211:
----------------------------------------

I've played a little bit around this and I've created/amended the following 
classes:

- 
[CamelController|https://github.com/lburgazzoli/apache-camel/blob/d394fce4a6dc0be831bbdab0d256d9270aa0a8cc/platforms/commands/commands-core/src/main/java/org/apache/camel/commands/CamelController.java]
- 
[AbstractCamelControlle|https://github.com/lburgazzoli/apache-camel/blob/d394fce4a6dc0be831bbdab0d256d9270aa0a8cc/platforms/commands/commands-core/src/main/java/org/apache/camel/commands/AbstractCamelController.java]
- 
[CatalogComponentInfoCommand|https://github.com/lburgazzoli/apache-camel/blob/d394fce4a6dc0be831bbdab0d256d9270aa0a8cc/platforms/commands/commands-core/src/main/java/org/apache/camel/commands/CatalogComponentInfoCommand.java]

Running the command to get information about hdfs component it would produce:

{code}
HDFS :: For reading/writing from/to an HDFS filesystem using Hadoop 1.x.
------------------------------------------------------------------------

label: hadoop,file
maven: org.apache.camel/camel-hdfs/2.17-SNAPSHOT

componentProperties

Key                      Description
---                      -----------
jAASConfiguration        To use the given configuration for security with JAAS.

properties

Key                      Description
---                      -----------
hostName                 HDFS host to use
port                     HDFS port to use
path                     The directory path to use
blockSize                The size of the HDFS blocks
bufferSize               The buffer size used by HDFS
checkIdleInterval        How often (time in millis) in to run the idle checker 
background task. This option is only in use if the splitter strategy is IDLE.
chunkSize                When reading a normal file this is split into chunks 
producing a message per chunk.
compressionCodec         The compression codec to use
compressionType          The compression type to use (is default not in use)
connectOnStartup         Whether to connect to the HDFS file system on starting 
the producer/consumer. If false then the connection is created on-demand. 
Notice that HDFS may take up till 15 minutes to establish a connection as it 
has hardcoded 45 x 20 sec redelivery. By setting this option to false allows 
your application to startup and not block for up till 15 minutes.
fileSystemType           Set to LOCAL to not use HDFS but local java.io.File 
instead.
fileType                 The file type to use. For more details see Hadoop HDFS 
documentation about the various files types.
keyType                  The type for the key in case of sequence or map files.
openedSuffix             When a file is opened for reading/writing the file is 
renamed with this suffix to avoid to read it during the writing phase.
owner                    The file owner must match this owner for the consumer 
to pickup the file. Otherwise the file is skipped.
readSuffix               Once the file has been read is renamed with this 
suffix to avoid to read it again.
replication              The HDFS replication factor
splitStrategy            In the current version of Hadoop opening a file in 
append mode is disabled since it's not very reliable. So for the moment it's 
only possible to create new files. The Camel HDFS endpoint tries to solve this 
problem in this way: If the split strategy option has been defined the hdfs 
path will be used as a directory and files will be created using the configured 
UuidGenerator. Every time a splitting condition is met a new file is created. 
The splitStrategy option is defined as a string with the following syntax: 
splitStrategy=ST:valueST:value... where ST can be: BYTES a new file is created 
and the old is closed when the number of written bytes is more than value 
MESSAGES a new file is created and the old is closed when the number of written 
messages is more than value IDLE a new file is created and the old is closed 
when no writing happened in the last value milliseconds
valueType                The type for the key in case of sequence or map files
bridgeErrorHandler       Allows for bridging the consumer to the Camel routing 
Error Handler which mean any exceptions occurred while the consumer is trying 
to pickup incoming messages or the likes will now be processed as a message and 
handled by the routing Error Handler. By default the consumer will use the 
org.apache.camel.spi.ExceptionHandler to deal with exceptions that will be 
logged at WARN/ERROR level and ignored.
delay                    The interval (milliseconds) between the directory 
scans.
initialDelay             For the consumer how much to wait (milliseconds) 
before to start scanning the directory.
pattern                  The pattern used for scanning the directory
sendEmptyMessageWhenIdle If the polling consumer did not poll any files you can 
enable this option to send an empty message (no body) instead.
exceptionHandler         To let the consumer use a custom ExceptionHandler. 
Notice if the option bridgeErrorHandler is enabled then this options is not in 
use. By default the consumer will deal with exceptions that will be logged at 
WARN/ERROR level and ignored.
pollStrategy             A pluggable 
org.apache.camel.PollingConsumerPollingStrategy allowing you to provide your 
custom implementation to control error handling usually occurred during the 
poll operation before an Exchange have been created and being routed in Camel.
append                   Append to existing file. Notice that not all HDFS file 
systems support the append option.
overwrite                Whether to overwrite existing files with the same name
exchangePattern          Sets the default exchange pattern when creating an 
exchange
synchronous              Sets whether synchronous processing should be strictly 
used or Camel is allowed to use asynchronous processing (if supported).
backoffErrorThreshold    The number of subsequent error polls (failed due some 
error) that should happen before the backoffMultipler should kick-in.
backoffIdleThreshold     The number of subsequent idle polls that should happen 
before the backoffMultipler should kick-in.
backoffMultiplier        To let the scheduled polling consumer backoff if there 
has been a number of subsequent idles/errors in a row. The multiplier is then 
the number of polls that will be skipped before the next actual attempt is 
happening again. When this option is in use then backoffIdleThreshold and/or 
backoffErrorThreshold must also be configured.
greedy                   If greedy is enabled then the ScheduledPollConsumer 
will run immediately again if the previous run polled 1 or more messages.
runLoggingLevel          The consumer logs a start/complete log line when it 
polls. This option allows you to configure the logging level for that.
scheduledExecutorService Allows for configuring a custom/shared thread pool to 
use for the consumer. By default each consumer has its own single threaded 
thread pool.
scheduler                To use a cron scheduler from either camel-spring or 
camel-quartz2 component
schedulerProperties      To configure additional properties when using a custom 
scheduler or any of the Quartz2 Spring based scheduler.
startScheduler           Whether the scheduler should be auto started.
timeUnit                 Time unit for initialDelay and delay options.
useFixedDelay            Controls if fixed delay or fixed rate is used. See 
ScheduledExecutorService in JDK for details.
{code}

Is that what you'd expect ?


> Camel commands - camel-component-info
> -------------------------------------
>
>                 Key: CAMEL-8211
>                 URL: https://issues.apache.org/jira/browse/CAMEL-8211
>             Project: Camel
>          Issue Type: New Feature
>          Components: tooling
>    Affects Versions: 2.15.0
>            Reporter: Claus Ibsen
>            Priority: Minor
>             Fix For: Future
>
>
> A new camel-catalog-component-info command to display detailed information 
> about the component.
> We should show
> component description
> label(s)
> maven coordinate
> list of all its options and description for those
> This allows users to use these commands in tooling to read the component 
> documentation.
> In the future we may slurp in any readme.md files we have in the components 
> so we can do all component documentation in the source code and not use the 
> confluence wiki which gets out of sync etc.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to