Hi Andrew, okay, we'll try that and will report back when/if we get it working.
Cheers, Lars On Thu, Oct 20, 2022 at 2:29 AM Andrew Purtell <apurt...@apache.org> wrote: > No, that is insufficient. HBase must be recompiled against Hadoop 3 first > > cd /path/to/hbase > mvn clean install assembly:single -DskipTests -Dhadoop.profile=3.0 > -Dhadoop-three.version=XXX > > Then once the results are in your local maven cache or nexus instance, you > can compile Spark as indicated. > > > On Tue, Oct 18, 2022 at 11:39 PM Lars Francke <lars.fran...@gmail.com> > wrote: > > > Hi Andrew, > > > > thanks for the reply. > > I should have been more specific: We only tried to compile the "client" > > part that's used in Spark itself and we used the proper versions > > > > mvn -Dspark.version=XXX -Dscala.version=XXX -Dhadoop-three.version=XXX > > -Dscala.binary.version=XXX -Dhbase.version=XXX clean package > > > > I assume that should pull in the correct dependencies but I have to admit > > that I didn't check, took it straight from the readme. > > We wanted to try the server bit for the RegionServers afterwards but > didn't > > even get to it yet. > > > > We have this on our radar though and might try to work through those > issues > > at some point. > > If we get started on that I'll ping the list. > > > > Cheers, > > Lars > > > > On Wed, Oct 19, 2022 at 1:41 AM Andrew Purtell <apurt...@apache.org> > > wrote: > > > > > Out of the box use is going to be problematic without recompiling HBase > > for > > > Hadoop 3. Spark 3.3 ships with Hadoop 3.3.2. Apache HBase 2.4.x (and > all > > > 2.x) releases are compiled against Hadoop 2. Link errors > (ClassNotFound, > > > NoClassDef, etc) I think are to be expected because the class > hierarchies > > > of various Hadoop things have been incompatibly changed in 3.x releases > > > relative to 2.x. This is not unreasonable. Semantic versioning suggests > > > breaking changes can be expected in a major version increment. > > > > > > Users probably need to do a holistic (or hermetic, if you prefer) build > > of > > > their bill of materials before testing it or certainly before shipping > > it. > > > Build your HBase for the version of Hadoop you are actually shipping it > > > with, as opposed to whatever the upstream project picks as a default > > build > > > target. They are called "convenience binaries" by the project and the > > > Foundation for a reason. Convenience may vary according to your > > > circumstances. When HBase finally ships builds compiled against Hadoop > 3 > > by > > > default, anyone still using 2.x in production will face the same > problem > > > (in reverse). The Phoenix project also faces this issue for what it's > > > worth. Their readme and build instructions walk users through > rebuilding > > > HBase using -Dhadoop.profile=3.0 as a first step as well. > > > > > > > > > On Mon, Oct 17, 2022 at 1:52 PM Lars Francke <lars.fran...@gmail.com> > > > wrote: > > > > > > > Hi everyone, > > > > > > > > we've just recently tried getting the HBase Spark connector running > > > against > > > > Spark 3.3 and HBase 2.4.x and failed miserably. It was a mess of > Scala > > > and > > > > Java issues, classpath, NoClassDef etc. > > > > > > > > The trauma is too recent for me to dig up the details but if needed I > > can > > > > ;-) > > > > > > > > For now I'm just wondering if anyone has succeeded using this > > > combination? > > > > > > > > Cheers, > > > > Lars > > > > > > > > > > > > > -- > > > Best regards, > > > Andrew > > > > > > Unrest, ignorance distilled, nihilistic imbeciles - > > > It's what we’ve earned > > > Welcome, apocalypse, what’s taken you so long? > > > Bring us the fitting end that we’ve been counting on > > > - A23, Welcome, Apocalypse > > > > > > > > -- > Best regards, > Andrew > > Unrest, ignorance distilled, nihilistic imbeciles - > It's what we’ve earned > Welcome, apocalypse, what’s taken you so long? > Bring us the fitting end that we’ve been counting on > - A23, Welcome, Apocalypse >