Hi Scott,

please read http://docs.graylog.org/en/2.0/pages/configuring_webif.html. 
The Graylog REST API and the web interface both need to be accessible by 
your web browser.

Cheers,
Jochen

On Friday, 3 June 2016 14:15:31 UTC+2, Scott John wrote:
>
> Thanks for the Help that seems to have fixed the connection to 
> elasticsearch but the server is still not coming up.  The website says 
> server connection unavailable and this is the only error in the log and 
> that error did not reappear after I restarted the service.  I have tried it 
> with 
> rest_listen_uri = http://127.0.0.1:12900/
> both set to 127.0.0.1 and also the server's actual IP
>
> 2016-06-03T07:49:05.021-04:00 ERROR [KafkaJournal] Read offset 243327505 
> before start of log at 245059183, starting to read from the beginning of 
> the journal.
> 2016-06-03T07:49:12.818-04:00 INFO  [connection] Opened connection 
> [connectionId{localValue:9, serverValue:79}] to localhost:27017
> 2016-06-03T07:49:12.819-04:00 INFO  [connection] Opened connection 
> [connectionId{localValue:8, serverValue:78}] to localhost:27017
> 2016-06-03T07:50:26.705-04:00 INFO  [connection] Opened connection 
> [connectionId{localValue:11, serverValue:82}] to localhost:27017
> 2016-06-03T07:50:26.706-04:00 INFO  [connection] Opened connection 
> [connectionId{localValue:12, serverValue:81}] to localhost:27017
> 2016-06-03T07:50:26.713-04:00 INFO  [connection] Opened connection 
> [connectionId{localValue:10, serverValue:80}] to localhost:27017
>
> Is there any other default values that i may need to adjust with this 
> having used to be a 1.3 system?
>
> Thanks,
> Scott
>
>
> On Friday, June 3, 2016 at 3:42:27 AM UTC-4, Jochen Schalanda wrote:
>>
>> Hi Scott,
>>
>> the default value for elasticsearch_cluster_name is "graylog" (see 
>> https://github.com/Graylog2/graylog2-server/blob/2.0.2/misc/graylog.conf#L210-L213)
>>  
>> but your Elasticsearch node is using "graylog2". You have to use the exact 
>> same cluster name in both configuration files.
>>
>> Cheers,
>> Jochen
>>
>> On Thursday, 2 June 2016 22:37:24 UTC+2, Scott John wrote:
>>>
>>> I am getting the following error in my graylog log file.
>>>
>>> 2016-06-02T16:16:55.327-04:00 WARN  [BlockingBatchedESOutput] Error 
>>> while waiting for healthy Elasticsearch cluster. Not flushing.
>>> java.util.concurrent.TimeoutException: Elasticsearch cluster didn't get 
>>> healthy within timeout
>>>         at 
>>> org.graylog2.indexer.cluster.Cluster.waitForConnectedAndHealthy(Cluster.java:179)
>>>  
>>> ~[graylog.jar:?]
>>>         at 
>>> org.graylog2.indexer.cluster.Cluster.waitForConnectedAndHealthy(Cluster.java:184)
>>>  
>>> ~[graylog.jar:?]
>>>         at 
>>> org.graylog2.outputs.BlockingBatchedESOutput.flush(BlockingBatchedESOutput.java:112)
>>>  
>>> [graylog.jar:?]
>>>         at 
>>> org.graylog2.outputs.BlockingBatchedESOutput.write(BlockingBatchedESOutput.java:105)
>>>  
>>> [graylog.jar:?]
>>>         at 
>>> org.graylog2.buffers.processors.OutputBufferProcessor$1.run(OutputBufferProcessor.java:189)
>>>  
>>> [graylog.jar:?]
>>>         at 
>>> com.codahale.metrics.InstrumentedExecutorService$InstrumentedRunnable.run(InstrumentedExecutorService.java:176)
>>>  
>>> [graylog.jar:?]
>>>         at 
>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) 
>>> [?:1.8.0_91]
>>>         at java.util.concurrent.FutureTask.run(FutureTask.java:266) 
>>> [?:1.8.0_91]
>>>         at 
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>  
>>> [?:1.8.0_91]
>>>         at 
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>  
>>> [?:1.8.0_91]
>>>         at java.lang.Thread.run(Thread.java:745) [?:1.8.0_91]
>>>
>>> The web interface comes up but reports that it cannot connect to the 
>>> server on port 12900.
>>> GL - 2.0.2
>>> ES - 2.3.3
>>> MDB - 3.0.12
>>> OS - RHEL 7.2
>>> Installed from repos.
>>>
>>> curl -XGET http://localhost:9200/_cluster/health?pretty
>>> {
>>>   "cluster_name" : "graylog2",
>>>   "status" : "green",
>>>   "timed_out" : false,
>>>   "number_of_nodes" : 1,
>>>   "number_of_data_nodes" : 1,
>>>   "active_primary_shards" : 48,
>>>   "active_shards" : 48,
>>>   "relocating_shards" : 0,
>>>   "initializing_shards" : 0,
>>>   "unassigned_shards" : 0,
>>>   "delayed_unassigned_shards" : 0,
>>>   "number_of_pending_tasks" : 0,
>>>   "number_of_in_flight_fetch" : 0,
>>>   "task_max_waiting_in_queue_millis" : 0,
>>>   "active_shards_percent_as_number" : 100.0
>>> }
>>>
>>> elasticsearch.yml
>>>
>>> cluster.name: graylog2
>>> network.host: 127.0.0.1 
>>> discovery.zen.ping.multicast.enabled: false
>>> discovery.zen.ping.unicast.hosts: ["127.0.0.1:9300"]
>>> script.inline: on
>>> script.indexed: on
>>>
>>> server.conf
>>>
>>> is_master = true
>>> node_id_file = /etc/graylog/server/node-id
>>> password_secret = XXXX
>>> root_password_sha2 = XXXX
>>> root_email = "[email protected]"
>>> root_timezone = UTC
>>> rest_listen_uri = http://127.0.0.1:12900/
>>> rotation_strategy = count
>>> elasticsearch_max_docs_per_index = 20000000
>>> elasticsearch_max_number_of_indices = 20
>>> retention_strategy = delete
>>> elasticsearch_shards = 4
>>> elasticsearch_replicas = 0
>>> elasticsearch_index_prefix = graylog2
>>> allow_leading_wildcard_searches = false
>>> allow_highlighting = false
>>> elasticsearch_http_enabled = false
>>> elasticsearch_discovery_zen_ping_multicast_enabled = false
>>> elasticsearch_discovery_zen_ping_unicast_hosts = 127.0.0.1:9300
>>> elasticsearch_network_host = 127.0.0.1
>>> elasticsearch_analyzer = standard
>>> output_batch_size = 500
>>> output_flush_interval = 1
>>> output_fault_count_threshold = 5
>>> output_fault_penalty_seconds = 30
>>> processbuffer_processors = 5
>>> outputbuffer_processors = 3
>>> processor_wait_strategy = blocking
>>> ring_size = 65536
>>> inputbuffer_ring_size = 65536
>>> inputbuffer_processors = 2
>>> inputbuffer_wait_strategy = blocking
>>> message_journal_enabled = true
>>> message_journal_dir = /var/lib/graylog-server/journal
>>> dead_letters_enabled = false
>>> lb_recognition_period_seconds = 3
>>> mongodb_uri = mongodb://localhost/graylog2
>>> mongodb_max_connections = 100
>>> mongodb_threads_allowed_to_block_multiplier = 5
>>> mongodb_useauth = false
>>> transport_email_enabled = false
>>> transport_email_hostname = mailhost.email.com
>>> transport_email_port = 587
>>> transport_email_use_auth = false
>>> transport_email_use_tls = true
>>> transport_email_subject_prefix = [graylog2]
>>> transport_email_from_email = [email protected]
>>> transport_email_web_interface_url = https://glog.email.com
>>> content_packs_dir = /usr/share/graylog-server/contentpacks
>>> content_packs_auto_load = grok-patterns.json
>>>
>>>
>>> What am I missing here?
>>>
>>

-- 
You received this message because you are subscribed to the Google Groups 
"Graylog Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/graylog2/e29d16fb-fa28-4c62-b1f0-ba75e577a74e%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to