Re: File corruption after power cut

2019-09-04 Thread Fred Dushin
Just remember if you are not running AAE, you will want to do a manual repair on that partition, in order to ensure you have adequate replication: https://docs.riak.com/riak/kv/2.2.3/using/repair-recovery/repairs.1.html#repairing-partitions > On Sep 4, 2019, at 5:19 AM, Bryan Hunt > wrote: >

Re: [ANN] Riak 2.9.0 - Release Candidate 1 Available

2019-02-01 Thread Fred Dushin
o additional > resources to play with. > > Nicholas > > From: riak-users On Behalf Of Fred Dushin > Sent: 01 February 2019 22:47 > To: riak-users@lists.basho.com > Subject: Re: [ANN] Riak 2.9.0 - Release Candidate 1 Available > > Wouldn't it be better to do a `riak-admi

Re: [ANN] Riak 2.9.0 - Release Candidate 1 Available

2019-02-01 Thread Fred Dushin
Wouldn't it be better to do a `riak-admin replace`? Leave could be problematic if there are other nodes in the cluster that are under-provisioned (disk space, for example). Plus a leave and add would move the data around the cluster twice, for each node in the cluster, whereas a replace would

Re: High disk usage on node

2018-11-01 Thread Fred Dushin
I think your best bet is to do a force-replace (and then a manual repair, if you are not using AAE) with a node that has higher capacity than your current standby. You are correct that replacing with your standby will fail when you run repairs and end up running out of space. I think you do

Re: rolling upgrades riak kv OSS 2.14 to 2.2.3

2018-07-09 Thread Fred Dushin
There are some changes to the format of AAE tree data, if you have AAE enabled. IIRC, AAE tree upgrades will not kick in until all nodes in the cluster are upgraded to 2.2 or later, but they will kick in automatically, once full upgrade is detected. You can disable the new AAE format, but

Re: Riak-2.2.5 progress update

2018-03-27 Thread Fred Dushin
@russeldb do you have ordained values for `current`, `previous`, and `legacy` to test against? Always the :bane: of riak_test -Fred > On Mar 27, 2018, at 1:47 PM, Russell Brown wrote: > > Giddyup died when basho was shuttered. These test runs have all been on > private

Re: RIAK Roadmap Workshop

2017-09-22 Thread Fred Dushin
Thank you, Andrew, and Bet365 for hosting this event, and for shepherding Riak through these difficult times. I will not be able to attend, but under enhancements, I would encourage the group to discuss the late enhancements to Riak/KV to support efficient bucket and key listing for sorted

Re: Decipher Solr error messages

2017-07-01 Thread Fred Dushin
=true; And of course you should be able to use the full query language of Solr to return the result set you desire (bucket type, bucket name, key, etc). -Fred > On Jul 1, 2017, at 10:30 AM, Fred Dushin <f...@dushin.net> wrote: > > My guess is that you have data that is either not UTF-8 e

Re: Decipher Solr error messages

2017-07-01 Thread Fred Dushin
My guess is that you have data that is either not UTF-8 encoded or is failing to decode into a JSON payload (peeking quickly at the mochi_json code, there are a lot of case statements, as you would expect in a parser). There is a call to error logging on line 165 in Riak 2.2.2, so I assume you

Re: Riak getting very slow

2017-06-28 Thread Fred Dushin
When you say "All not have solr on", do you mean not all nodes have search enabled? If you are measuring Solr index latencies, then you definitely have Solr on at least one node. Or is this just a typo? Going on the assumption you have search enabled on all nodes (you should, if you are

Re: Stopping solrq supervisor in log; What is this?

2017-06-22 Thread Fred Dushin
This might be a relatively benign message, which might be better at debug scope than info scope. This would have been introduced in 2.0.8 or 2.2.2 or later. Essentially, there are solq supervisors for each partition/index pair (for each partition on a node), and when these supervisors start or

Re: Solr search response time spikes

2017-06-22 Thread Fred Dushin
It's pretty strange that you are seeing no search latency measurements on node 5. Are you sure your round robining is working? Are you favoring node 1? In general, I don't think which node you hit for query should make a difference, but I'd have to stare at the code some to be sure. In

Re: Riak getting very slow

2017-06-17 Thread Fred Dushin
Hi Amol, When you say "All not have solr on", do you mean not all nodes have search enabled? If you are measuring Solr index latencies, then you definitely have Solr on at least one node. Or is this just a typo? Going on the assumption you have search enabled on all nodes (you should, if

Re: Riak issue after power outage

2017-06-01 Thread Fred Dushin
There may be better ways, but what I would do is: Stop Riak, blow away your ${platform_data_dir}/anti_entropy directory (and yz_anti_entropy, if you have search enabled), and restart. This will force a rebuild of your hash trees. They rebuild automatically anyway; this is just making them

Re: Search index partitioning in Riak-KV

2017-05-19 Thread Fred Dushin
There is a Solr instance per Riak node, and each Solr instance contains a Solr core for the Riak index created. Every replica of a Riak k/v pair has a corresponding document in the Solr instance on the same node as the vnode that stores the replica (in fact for each sibling, if your objects

Re: How Riak Handle Request?

2017-04-17 Thread Fred Dushin
What do you mean by 'process'? If you mean operating system process, then, No. A Riak instance runs in a single OS process. There is a Riak instance per "node" in the cluster, where a "node" is a machine, container, etc. Probably what the OMG guys used to call a "capsule" (cute) There are

Re: anti_entropy and solr taking up suspiciously large amounts of space

2017-04-05 Thread Fred Dushin
Batching parameters only affect the update operations, so while deletes are relevant, none of the batching parameters will affect the issue you are seeing with TTLs that are only implemented in the back end. One thing that might help, in the future, is the sweeper functionality, which is

Re: Problem with Solr Search

2017-04-01 Thread Fred Dushin
> On Mar 30, 2017, at 12:58 AM, Nguyễn Văn Nghĩa Em wrote: > > I have met some problems: > 1. Create search index with error: "Can not create search index on all > nodes". This error appear after 30 seconds timeout. I opened the index folder > in /var/lib/riak/yz, I found

Re: Stale index entries after delete

2017-03-17 Thread Fred Dushin
Hello Bernard, Thank you for writing this up, and apologies for the delay in responding. I have not been able to readily reproduce this in my test environment, and our Riak tests (e.g., [1]) test scenarios very much like this. If this is a test environment, can you try stopping riak, deleting

Re: Update solr schema

2017-03-06 Thread Fred Dushin
Hi Alexander. Re-indexing is still required if you modify schema. Data stored in Riak is the "database of record", and the data stored in Solr is secondary. -Fred > On Mar 6, 2017, at 5:52 AM, Alexander Popov wrote: > > After release riak 2.2 does it any new

Re: Yokozuna's inconsistent query problem

2017-02-23 Thread Fred Dushin
Hello Witeman, What you are seeing with your two queries is the result of two different coverage plans, querying different parts of the cluster. Riak Search translates coverage plans to Solr sharded queries, and will periodically change the coverage plan, so as to more evenly distribute

Re: Riak: reliable object deletion

2017-02-23 Thread Fred Dushin
Running a Solr query has no impact on writes -- Riak search queries are direct pass throughs to Solr query and don't touch any of the salient Riak systems (batching writes to Solr, YZ AAE, etc). I believe the timing of the reappearance is a coincidence. Is it possible the object reappeared

Re: Monitor Riak Network Port and IO Port

2016-12-08 Thread Fred Dushin
The process is (typically) beam.smp, though you may have multiple on your machine, if for example, you are connected to riak via the console, or if you are running administrative commands (e.g., riak-admin). For the ports (if that is also what you are looking for) see:

Re: How to specify dismax related parameters like qf

2016-10-17 Thread Fred Dushin
The internal solr API will not use the distributed queries generated from coverage plans. You will only get results from the local node. Theoretically, you could aggregate and de-duplicate across multiple nodes, but that would result in more data movement than necessary, as it does not

Re: Using cursorMark with '_yz_rk'

2016-09-21 Thread Fred Dushin
Okay, I probably spoke too soon. While Solr 4.7 supports cursor marks, we do have an issue in Riak (or Yokozuna) whereby it is actually impractical to use cursor marks for query. The problem is that while Yokozuna uses coverage plans generate a filter query that will guarantee that we get no

Re: Solr search performance

2016-09-19 Thread Fred Dushin
All great questions, Sean. A few things. First off, for result sets that are that large, you are probably going to want to use Solr cursor marks [1], which are supported in the current version of Solr we ship. Riak allows queries using cursor marks through the HTTP interface. At present, it

Re: Riak 2.1.3 - Multiple indexes created by Solr for the same Riak object

2016-09-12 Thread Fred Dushin
Hi Weixi, You might have to try describing your use case in more detail. Solr Indices are independent from Riak objects. They are, instead, associated with riak buckets (or bucket types), and an object (key/value) can only be associated with one bucket. Therefore, a Riak object can only be

Re: Riak search, post schema change reindexation

2016-08-29 Thread Fred Dushin
Hi Guillame, A few questions. What version of Riak? Does the reindexing need to occur across the entire cluster, or just on one node? What are the expectations about query-ability while re-indexing is going on? If you can afford to take a node out of commission for query, then one approach

Re: Querying SOLR outside of Riak

2016-05-16 Thread Fred Dushin
Hi Alex, Other people have chimed in, but let me repeat that while the internal_solr interface is accessible via HTTP (and needs to be, at least from Riak processes), you cannot use that interface to query Solr and expect a correct result set (unless you are using a single node cluster with an

Re: Very slow acquisition time (99 percentile) while fast median times

2016-05-03 Thread Fred Dushin
Hi Guillaume, From your bucket properties it looks like you are using search, and I assume that is search 2.0 (i.e., yokozuna), and not the legacy Riak Search. It is true that in the current 2.0 and 2.1 trunks the indexing into Solr via Yokozuna is synchronous with the vnode -- very long times

Re: Riak server crashing

2016-05-02 Thread Fred Dushin
Acceptor not being able to create a server side socket smells like you've run out of file descriptors. > On May 2, 2016, at 10:25 AM, Ricardo Mayerhofer wrote: > > Hi all, > I've a riak server that was running fine for 1 month, now it keep crashing. > Restart has no

Re: Solr http endpoint and POST

2016-04-26 Thread Fred Dushin
Hi Alexander, Do you mean, is it possible to run a Yokozuna query through Riak using a HTTP POST, instead of a GET? Have you tried the pattern used here? https://github.com/basho/yokozuna/blob/2.1.2/riak_test/yz_search_http.erl#L68

Re: solrconfig.xml in riak

2016-04-26 Thread Fred Dushin
Hi Fasil, Riak does not provide any administrative operations for this file. You can, however, edit the file for each Solr core on each riak node. You will find it under: /yz//conf You will need to restart either riak on that node, the Solr on that node, or the Yokozuna application on that

Re: How to index List in riak so that it will available for solr

2016-04-22 Thread Fred Dushin
Hi Anil, Thanks for your question. Riak uses the HTTP interface for inserting data into Solr, not Solrj. We do not therefore support indexing POJOs -- they must be translated to a form that is consumable by the HTTP interface (internally we use JSON, but Riak natively supports several

Re: Solr error message

2016-04-11 Thread Fred Dushin
Hi Jim, Interesting problem. That error is occurring here: https://github.com/basho/yokozuna/blob/2.1.2/src/yz_cover.erl#L275 because length(Mapping) and length(UniqNodes) are unequal: https://github.com/basho/yokozuna/blob/2.1.2/src/yz_cover.erl#L262 This might be because you are getting

Re: Yokozuna inconsistent search results

2016-04-05 Thread Fred Dushin
Hi Oleksiy, I assume you are getting this information through riak-admin. Can you post the results here? If you want to dig deeper, you can probe the individual hash trees for their build time. I will paste a few snippets of erlang here, which I am hoping you can extend to use with list

Re: Yokozuna inconsistent search results

2016-03-11 Thread Fred Dushin
Hi Oleksiy, This is definitely pointing to an issue either in the coverage plan (which determines the distributed query you are seeing) or in the data you have in Solr. I am wondering if it is possible that you have some data in Solr that is causing the rebuild of the YZ AAE tree to

Re: Yokozuna inconsistent search results

2016-02-26 Thread Fred Dushin
I would check the coverage plans that are being used for the different queries, which you can usually see in the headers of the resulting document. When you run a search query though yokozuna, it will use a coverage plan from riak core to find a minimal set of nodes (and partitions) to query

Re: Secondary indexes do not work with leveldb and write_once?

2016-01-13 Thread Fred Dushin
Hi Eugene, Unfortunately, the write-once put path does not support pre-commit hooks, which the 2i system uses for indexing. So secondary indexing is not currently supported with write-once. However, the indexing through yokozuna (Riak Search 2.0) is done in the vnode back-end, and indexing

Re: High CPU on a single node in production

2016-01-06 Thread Fred Dushin
Hi Josh, Sorry for not getting back sooner. I am not entirely sure what is going on with your handoffs. It could be that you have overloaded Solr with handoff activity, and that is causing vnodes to become unresponsive. We are actively working on a fix for this, which allows vnodes to

Re: Solr indexes are dropped after recovering a node

2015-12-24 Thread Fred Dushin
Hi Istvan, A couple of questions: 1. Do the buckets (or bucket types) the data is written to still contain the search index to be used to index the data? 2. Did the indices get re-created on the new nodes you added? You can verify this by looking at /var/lib/riak/yz/ (or wherever your

Re: Performance with riak can't reach expected

2015-10-07 Thread Fred Dushin
Hi Yang, Could you say a little more about what your requirements are, particularly around reliability? For example, what is your n_val, if you are using a 2-node cluster? It would also help to know how many worker processes you have in your basho bench config (the concurrent setting), as

Re: Yokozuna indexes too slow

2015-10-02 Thread Fred Dushin
.com/basho/yokozuna/issues/320> > > I try to set > maxThreads to 150 > Acceptors to 10 > lowResourcesMaxIdleTime to 5 > in /usr/lib/riak/lib/yokozuna/priv/solr/etc/jetty.xml as recommended in > https://github.com/basho/yokozuna/issues/330 > <https://github.com/basho/yo

Re: Yokozuna indexes too slow

2015-10-01 Thread Fred Dushin
Is there any more information in these logs that you can share? For example, is this the only entry with this exception? Or are there more? Are there any associated stack traces? An EOF exception can come from many different scenarios. Is there anything in the Riak console.log that looks

Re: riak restart does not restart yokozuka well

2015-09-14 Thread Fred Dushin
Check the solr logs to see why it failed to shut down. If necessary, find the pid bound to port 8985 and kill it. -Fred > On Sep 14, 2015, at 5:28 AM, Alexander Popov wrote: > > i'm doing riak restart > and got 'ok' in answer, but node appears in shutdown state. > > in

Re: What's the maximum seconds to set index on bucket after creating the index

2015-09-02 Thread Fred Dushin
<http://basho.github.io/riak-erlang-client/> The order of precedence is: 1. client-defined 2. riak config 3. default (45 seconds) -Fred > On Sep 2, 2015, at 8:13 AM, Fred Dushin <fdus...@basho.com> wrote: > > What is the return value you are getting from > rick_pb_socket:cr

Re: What's the maximum seconds to set index on bucket after creating the index

2015-09-02 Thread Fred Dushin
What is the return value you are getting from rick_pb_socket:create_search_index? If it's ok, then the Solr cores should have been created on all nodes. Otherwise, you should check the logs for timeout messages, e.g., https://github.com/basho/yokozuna/blob/2.1.1/src/yz_index.erl#L443

Re: riak 2.1.1 starting error in mac os 10.10

2015-07-30 Thread Fred Dushin
You're getting a runtime load error because your are likely not linking against the right C++ runtime (because you built with gcc). IIRC the GCC C++ runtime and the C++ runtime that ships with Darwin are not ABI compatible. Welcome to C++. You might try hacking your environment using

Re: Error while installing riak on ubuntu 12.04

2015-07-08 Thread Fred Dushin
Generally this means you built a version of Erlang that did not end up building the crypto application. This usually happens if the configure script does not find the OpenSSL header files and libraries. You probably need to install one of the openssl-dev packages on whatever flavor of Linux

Re: Feedback for GSoC project - RIak Destination for Syslog-ng

2015-05-05 Thread Fred Dushin
First off, Parth, this is a really exciting project, and I'm glad you're taking it on. As an SIEM refugee, I have a few questions about the proposal and a few thoughts about syslog, generally, and that may help you work out some of your thoughts about data types and how you plan to structure

Re: Errors running rebar in a the riak project, and compiling rebar from source

2015-01-30 Thread Fred Dushin
It looks like your Erlang/OTP install does not have the crypto module -- how was it initially installed (RPM? apt?)? I have seen this happen on recent versions of OS X with a manually built Erlang/OTP, as the erlang configure scripts can't locate the OpenSSL header files, so the crypto