Hello,
an unknown OpenCL error is an indication that maybe your OpenCL
installation isn't working correctly. Can you please verify that OpenCL
works fine (i.e. outside of ViennaCL)?
I'm not sure whether the problem has been resolved already - the
stackoverflow page suggests so.
Best
this as input for a better application next year :-)
Best regards,
Karli
On 2/7/23 18:37, Karl Rupp wrote:
Dear all,
thanks for all the input and help. Our application has been submitted,
let's keep our fingers crossed.
Also, this is a friendly reminder to fill out the details on the
GSoC-topics
are properly communicated. :-)
Thanks and best regards,
Karli
On 2/6/23 20:24, Karl Rupp wrote:
Hello all,
thanks for proposing projects. I've created the suggestions so far as
'issues' in the issue tracker on Gitlab, prefixed by 'GSoC:'. Please add
a better description to your suggestions so
ust async
* Differentiable programming using Enzyme with PETSc
I like all those.
Matt
Karl Rupp mailto:r...@iue.tuwien.ac.at>> writes:
> Dear PETSc developers,
>
> in order to attract students to PETSc development, I'm thinking
about a
> PETSc ap
Dear PETSc developers,
in order to attract students to PETSc development, I'm thinking about a
PETSc application for Google Summer of Code (GSoC) 2023:
https://summerofcode.withgoogle.com/programs/2023
The org application deadline is February 7, i.e. in 4 days. This
application is - roughly
in the examples.
The thing is I want to compute a QR (dgeqrf from Lapack) on the GPU, but
I could not find the right function for now.
Many thanks.
Regards,
Sandra
El jue, 2 dic 2021 a las 14:51, Karl Rupp (<mailto:r...@iue.tuwien.ac.at>>) escribió:
Hi Sandra,
ViennaCL is a header-onl
13:58, Karl Rupp (mailto:r...@iue.tuwien.ac.at>>) escribió:
Hello,
> They are the same compiler. Is that what you expected?
>
> jbelloch@odroid:~/EQU_FILTERING_sandra/ViennaCL-1.7.1$ c++
--version
> g++-4.8.real (Ubuntu/Linar
Hello,
They are the same compiler. Is that what you expected?
jbelloch@odroid:~/EQU_FILTERING_sandra/ViennaCL-1.7.1$ c++ --version
g++-4.8.real (Ubuntu/Linaro 4.8.2-19ubuntu1) 4.8.2
Copyright (C) 2013 Free Software Foundation, Inc.
This is free software; see the source for copying conditions.
a few scalars...
operator +=
operator *=
operator -=
operator /=
operator +
multiple operators
mixed operations
CPU scalar s3: 42
GPU scalar vcl_s3: 42
TUTORIAL COMPLETED SUCCESSFULLY
Regards,
Sandra
El vie, 26 nov 2021 a las 10:44, Karl Rupp (<mailto:r...@iue.tuwien.ac.at>
r work if you build and execute manually?
Best regards,
Karli
On 11/26/21 10:34, SANDRA CATALAN PALLARES wrote:
Hi Karli,
No problem. Here is the new output.
Regards,
Sandra
El vie, 26 nov 2021 a las 9:59, Karl Rupp (<mailto:r...@iue.tuwien.ac.at>>) escribió:
Dear Sandra,
sor
t;) escribió:
The output is attached.
Best regards,
Sandra
El vie, 19 nov 2021 a las 14:03, Karl Rupp (mailto:r...@iue.tuwien.ac.at>>) escribió:
Ok, so now we really need to actual compiler call from CMake. In
build/:
$> cmake .. -DC
l-support@lists.sourceforge.net> and supply at least the
following information:
* Operating System
* Which OpenCL implementation (AMD, NVIDIA, etc.)
* ViennaCL version
Many thanks in advance!
Aborted
Best regards,
Sandra
El vie, 19 nov 2021 a las 13:27, Karl Rupp (<mailto:r...@i
sandra/opencl
-DOPENCL_LIBRARY=/home/jbelloch/EQU_FILTERING_sandra/opencl/lib/libOpenCL.so
ENABLE_UBLAS . -DCMAKE_CXX_FLAGS="-DVIENNACL_DEBUG_ALL"/
/
/
Best regards,
Sandra
El vie, 19 nov 2021 a las 10:00, Karl Rupp (mailto:r...@iue.tuwien.ac.at>>) escribi
/=
operator +
multiple operators
mixed operations
CPU scalar s3: 42
GPU scalar vcl_s3: 42
TUTORIAL COMPLETED SUCCESSFULLY
Best regards,
Sandra
El vie, 19 nov 2021 a las 8:49, Karl Rupp (<mailto:r...@iue.tuwien.ac.at>>) escribió:
Indeed!
Line 58 causes the error:
https://g
istery continues:
jbelloch@odroid:~/EQU_FILTERING_sandra/test_opencl$ ./a.out
# Platforms found: 1
0
Best regards,
Sandra
El vie, 19 nov 2021 a las 8:33, Karl Rupp (<mailto:r...@iue.tuwien.ac.at>>) escribió:
Good morning,
oh - now I'm puzzled. :-/
Can you please add a
s
: 1
Cheers,
Sandra
El jue, 18 nov 2021 a las 19:48, Karl Rupp (<mailto:r...@iue.tuwien.ac.at>>) escribió:
thanks! This indicates that there is no OpenCL platform detected on
your
system. Basically this is the simplest query of all.
Can you try to compile and
t.c:89
#3 0xb6f517a8 in __gnu_cxx::__verbose_terminate_handler() () from
/usr/lib/arm-linux-gnueabihf/libstdc++.so.6
#4 0xb6f501c8 in ?? () from /usr/lib/arm-linux-gnueabihf/libstdc++.so.6
Backtrace stopped: previous frame identical to this frame (corrupt stack?)/
El jue, 1
f501c8 in ?? () from /usr/lib/arm-linux-gnueabihf/libstdc++.so.6
Backtrace stopped: previous frame identical to this frame (corrupt stack?)/
/
/
Regards,
Sandra
El jue, 18 nov 2021 a las 16:27, Karl Rupp (<mailto:r...@iue.tuwien.ac.at>>) escribió:
Hi,
thank you fo
dium=email_source=link_campaign=sig-email_content=webmail_term=icon>>
> Virus-free. www.avast.com <http://www.avast.com>
>
<https://www.avast.com/sig-email?utm_medium=email_source=link_campaign=sig-email_content=webmail_term=link
sig-email_content=webmail_term=icon>
Virus-free. www.avast.com
<https://www.avast.com/sig-email?utm_medium=email_source=link_campaign=sig-email_content=webmail_term=link>
<#DAB4FAD8-2DD7-40BB-A1B8-4E2AA1F9FDF2>
El vie, 12 nov 2021 a las 16:24, Karl Rupp (<mailto:r...@iue.tuwien.ac.at&
Hi,
this error usually indicates that your OpenCL installation isn't working
properly. Does clinfo work from the command line?
Best regards,
Karli
On 11/12/21 14:05, SANDRA CATALAN PALLARES wrote:
Good morning,
I am trying to run the examples included in Vienna, but I get the
following
Hello,
thank you for describing your application in great detail! With a system
size of 2000 to 5 unknowns you are most likely better off with
staying on the CPU (assuming that your system is indeed rather sparse
with less than about 100 nonzeros per row on average). This is because
each
Hi Bill,
if cmake worked out fine, then you find the executables in your build
folder under `examples/tutorial/`, e.g. `examples/tutorial/blas1`. Can
you please give me a short pointer on where you found the reference to
"examples/blas1” so that I can fix it?
Thanks and best regards,
Karl
Hi,
such adjustments should not need a direct modification of the theme. One
can just override the CSS settings in custom CSS files instead:
https://docs.readthedocs.io/en/stable/guides/adding-custom-css.html
The benefit of such an approach is that all future updates of the theme
will
Since we cannot post issues (reported here
https://forum.gitlab.com/t/creating-new-issue-gives-cannot-create-issue-getting-whoops-something-went-wrong-on-our-end/41966?u=bsmith)
here is my issue so I don't forget it.
I think
err = WaitForCUDA();CHKERRCUDA(err);
ierr =
Hi,
Since we cannot post issues (reported here
https://forum.gitlab.com/t/creating-new-issue-gives-cannot-create-issue-getting-whoops-something-went-wrong-on-our-end/41966?u=bsmith)
here is my issue so I don't forget it.
I think
err = WaitForCUDA();CHKERRCUDA(err);
ierr =
Hi Fernando,
did you verify that the OpenCL driver is installed properly on your
machine? Are you able to run any other OpenCL applications?
Best regards,
Karli
On 7/18/20 7:00 PM, Fernando García Sedano wrote:
Hi,
I am trying to install gpuR package, and we always get the error
From a practical standpoint it seems to me that NCCL is an offering to
a community that isn't used to MPI. It's categorized as 'Deep Learning
Software' on the NVIDIA page ;-)
The section 'NCCL and MPI' has some interesting bits:
Hi Jacob,
the recommendation in the past was to use MPICH as it is (was?)
valgrind-clean. Which MPI do you use? OpenMPI used to have these kinds
of issues. (My information might be outdated)
Best regards,
Karli
On 6/2/20 2:43 AM, Jacob Faibussowitsch wrote:
Hello All,
TL;DR: valgrind
Hi,
yes, Satish is right, this build is a CPU-build. Add `--with-opencl=1` :-)
Best regards,
Karli
On 4/16/20 12:31 AM, Satish Balay wrote:
From prior e-mail - you wanted to use AMD GPU on OSX. This build below is CPU
build - not for GPU. [Karl can confirm]
I think OSX has OpenCL installed
Hi,
yes, Satish is right, this build is a CPU-build. Add `--with-opencl=1` :-)
Best regards,
Karli
On 4/16/20 12:31 AM, Satish Balay wrote:
From prior e-mail - you wanted to use AMD GPU on OSX. This build below is CPU
build - not for GPU. [Karl can confirm]
I think OSX has OpenCL installed
Hi Abel,
try the ViennaCL backend. It used to work in the past, but I don't know
the current state of AMD drivers and OpenCL on OS X.
Best regards,
Karli
On 4/14/20 5:50 AM, huabel via petsc-users wrote:
Dear PETSc users,
Does PETSc suppurt AMD GPU card (such as RX 5700) on OS X?
The fluctuations in this example have been fixed a few months ago; the
issue was the use of multiple streams instead of a single one. Maybe
additional CUDA streams have been reintroduced recently?
Best regards,
Karli
On 4/2/20 5:02 AM, Junchao Zhang wrote:
I could not reproduce it locally.
Dear PETSc developers and PETSc users,
due to the recent Covid-19 outbreak in Europe there will not be a PETSc
User Meeting this year. We are looking into alternatives for keeping in
touch with our user base, e.g. via webinars. Suggestions welcome :-)
Thanks and best regards,
Karl
Dear PETSc developers and PETSc users,
due to the recent Covid-19 outbreak in Europe there will not be a PETSc
User Meeting this year. We are looking into alternatives for keeping in
touch with our user base, e.g. via webinars. Suggestions welcome :-)
Thanks and best regards,
Karl
CUDA
code). Thanks for the discussion at PP20.
On Sat, Feb 22, 2020 at 11:16 PM Karl Rupp <mailto:r...@iue.tuwien.ac.at>> wrote:
Hi Mark,
we are just finalizing the last few details (in particular: which
registration system to use) before sending out the anno
Hi Junchao,
I want to evaluate MatMult on GPU. I took a 2M x 2M matrix and ran with
6 mpi ranks and 6 GPUs. It took about 0.9 seconds.
How many nonzeros per row? With 0.9 seconds you should either have many
runs of MatMult, or a fairly dense matrix; or a really slow MatMult
kernel ;-)
Hi Mark,
we are just finalizing the last few details (in particular: which
registration system to use) before sending out the announcements and
putting the link on the main webpage. Just a matter of a few days. :-)
Best regards,
Karli
On 2/22/20 9:11 PM, Mark Adams wrote:
Maybe the
Hi Hong,
have you tried running the code through gprof and look at the output
(e.g. with kcachegrind)?
(apologies if this has been suggested already)
Best regards,
Karli
On 2/12/20 7:29 PM, Zhang, Hong via petsc-dev wrote:
On Feb 12, 2020, at 5:11 PM, Smith, Barry F. wrote:
ldd
Hi Rick,
have you verified that your matrices are positive definite? The problems
with 'nan' usually stem from the lack of positive definiteness or zeros
on the diagonal (as they often show up in saddle point problems).
The code snippet you provide looks fine. 200 are a lot of
Hi,
this looks like an issue with gpuR. Charles Determan is likely able to
help (in CC:).
Have you verified that your OpenCL SDK is installed properly, as the
error message suggests?
Best regards,
Karli
On 12/3/19 1:35 PM, Emir Smaka wrote:
esteemed,
trying to install the gpuR package
Hi,
Table 2 reports negative latencies. This doesn't look right to me ;-)
If it's the outcome of a parameter fit to the performance model, then
use a parameter name (e.g. alpha) instead of the term 'latency'.
Figure 11 has a very narrow range in the y-coordinate and thus
exaggerates the
to provide a good playground on time for the Summit
hackathon. In the meantime you can try the matrix market reader of
nsparse directly and see what you get, especially compared to cuSPARSE
and MKL.
Best regards,
Karli
Karl Rupp via petsc-dev writes:
Hi Richard,
CPU spGEMM is about
and compare against the other packages. In
the end it doesn't matter which package provides the best performance;
we just want to leverage it :-)
Best regards,
Karli
Karl Rupp via petsc-dev writes:
Hi Richard,
CPU spGEMM is about twice as fast even on the GPU-friendly case of a
single
Hi Richard,
CPU spGEMM is about twice as fast even on the GPU-friendly case of a
single rank: http://viennacl.sourceforge.net/viennacl-benchmarks-spmm.html
I agree that it would be good to have a GPU-MatMatMult for the sake of
experiments. Under these performance constraints it's not top
Hi Junchao,
I recall that Jed already suggested to make this a bitmask ~7 years ago ;-)
On the other hand: If we touch valid_GPU_array, then we should also use
a better name or refactor completely. Code like
(V->valid_GPU_array & PETSC_OFFLOAD_GPU)
simply isn't intuitive (nor does it make
Hi Mark,
OK, so now the problem has shifted somewhat in that it now manifests
itself on small cases. In earlier investigation I was drawn to
MatTranspose but had a hard time pinning it down. The bug seems more
stable now or you probably fixed what looks like all the other bugs.
I added
On Wed, Sep 25, 2019 at 5:26 AM Karl Rupp via petsc-dev
mailto:petsc-dev@mcs.anl.gov>> wrote:
On 9/25/19 11:12 AM, Mark Adams via petsc-dev wrote:
> I am using karlrupp/fix-cuda-streams, merged with master, and I
get this
> error:
>
> Could not execu
On 9/25/19 11:12 AM, Mark Adams via petsc-dev wrote:
I am using karlrupp/fix-cuda-streams, merged with master, and I get this
error:
Could not execute "['jsrun -g\\ 1 -c\\ 1 -a\\ 1 --oversubscribe -n 1
printenv']":
Error, invalid argument: 1
My branch mark/fix-cuda-with-gamg-pintocpu
ichard
On 9/23/19 8:09 PM, Karl Rupp wrote:
Hi,
`git grep cudaStreamCreate` reports that vectors, matrices and
scatters create their own streams. This will almost inevitably
create races (there is no synchronization mechanism implemented),
unless one calls WaitForGPU()
Hi,
`git grep cudaStreamCreate` reports that vectors, matrices and scatters
create their own streams. This will almost inevitably create races
(there is no synchronization mechanism implemented), unless one calls
WaitForGPU() after each operation. Some of the non-deterministic tests
can
On 9/22/19 6:15 AM, Jed Brown wrote:
Karl Rupp via petsc-dev writes:
Hi Junchao,
thanks, these numbers are interesting.
Do you have an easy way to evaluate the benefits of a CUDA-aware MPI vs.
a non-CUDA-aware MPI that still keeps the benefits of your
packing/unpacking routines?
I'd
Hi Junchao,
thanks, these numbers are interesting.
Do you have an easy way to evaluate the benefits of a CUDA-aware MPI vs.
a non-CUDA-aware MPI that still keeps the benefits of your
packing/unpacking routines?
I'd like to get a feeling of where the performance gains come from. Is
it due
Hi,
one way to test is to run a sequential example through nv-prof:
$> nvprof ./ex56 ...
https://devblogs.nvidia.com/cuda-pro-tip-nvprof-your-handy-universal-gpu-profiler/
If it uses the GPU, then you will get some information on the GPU
kernels called. If it doesn't use the GPU, the list
Hi all,
let me propose the following schedule for the next release:
* until Sunday, September 15: New pull requests are considered for the
upcoming release.
* from Monday, September 16, to Sunday, September 22: Fixing and merging
of open pull requests received by September 15. Extended
non-collective VecSet
12042c4bfa removing ViennaCL fix to GAMG
3c46958f6d fix bug with empty processor
8bcb2d50b7 fixed MPI lock from call to collective method
54cfeb1831 added missing settypes
9508265e8e adding support for MatTranspose
e5a6000419 adding fix for ViennaCL in MG
On Mon, 12 Aug
? (on vacation this week) I
will verify that master is all fixed if you get all my stuff integrated
when I get back to work in a week.
Thanks,
Mark
On Sat, Aug 3, 2019 at 10:50 AM Karl Rupp <mailto:r...@iue.tuwien.ac.at>> wrote:
If you ignore the initial ViennaCL-related commits and chec
fixed if you get all my stuff integrated
when I get back to work in a week.
Thanks,
Mark
On Sat, Aug 3, 2019 at 10:50 AM Karl Rupp <mailto:r...@iue.tuwien.ac.at>> wrote:
If you ignore the initial ViennaCL-related commits and check against
current master (that just received cher
If you ignore the initial ViennaCL-related commits and check against
current master (that just received cherry-picked updates from your PR),
then there are really only a few commits left that are not yet integrated.
(I'll extract two more PRs on Monday, so master will soon have your
fixes
Barry's branch. Or will that not work.
On Fri, Aug 2, 2019 at 12:03 PM Karl Rupp <mailto:r...@iue.tuwien.ac.at>> wrote:
FYI: The two branches are currently testing in `next-tmp` and are
likely
to be merged to master in ~5 hours.
Best regards,
Karli
On 8/2/1
FYI: The two branches are currently testing in `next-tmp` and are likely
to be merged to master in ~5 hours.
Best regards,
Karli
On 8/2/19 4:53 PM, Smith, Barry F. via petsc-dev wrote:
Yes, these are bugs in Stefano's work that got into master because we didn't
have comprehensive
Hi Mark,
feel free to submit a fresh pull request now. I looked at your latest
commit in the repository in order to cherry-pick it, but it looked like
it had a few other bits in it as well.
Best regards,
Karli
On 7/28/19 6:27 PM, Mark Adams via petsc-dev wrote:
This is looking good. I'm
On Fri, Jul 19, 2019 at 12:08 PM Karl Rupp <mailto:r...@iue.tuwien.ac.at>> wrote:
Hi Xiangdong,
> I can understand some of the numbers, but not the HtoD case.
>
> In DtoH1, it is the data movement from VecMDot. The size of data is
> 8.192KB, which
Hi Stefano,
I have just noticed we have different occurrences of the
valid_GPU_matrix flag in src/mat/interface and src/mat/utils
I think that how they are used now is wrong, as they assume that all
those operations can only be executed on the CPU, irrespective of the
specific type.
Is there
the exact source of these transfers, but
that is some effort. Let me know whether this is important information
for you, then I can do it.
Best regards,
Karli
Thank you.
Best,
Xiangdong
On Thu, Jul 18, 2019 at 1:14 PM Karl Rupp <mailto:r...@iue.tuwien.ac.at>> wrote:
Hi,
as you can see from the screenshot, the communication is merely for
scalars from the dot-products and/or norms. These are needed on the host
for the control flow and convergence checks and is true for any
iterative solver.
Best regards,
Karli
On 7/18/19 3:11 PM, Xiangdong via
That's just a manifestation of Satish merging really well today ;-)
Best regards,
Karli
On 5/30/19 1:11 AM, Smith, Barry F. via petsc-dev wrote:
I just got this same merged message sent to me three times.
In recent days I've received several sent to me twice.
It's not like we
Using alt files for testing is painful. Whenever you add, for example, a
new variable to be output in a viewer it changes the output files and you need
to regenerate the alt files for all the test configurations. Even though the
run behavior of the code hasn't changed.
I'm
Hi,
Scott and PETSc folks,
Using alt files for testing is painful. Whenever you add, for example, a
new variable to be output in a viewer it changes the output files and you need
to regenerate the alt files for all the test configurations. Even though the
run behavior of the code
Hi,
I fixed this warning after merge.
Best regards,
Karli
On 4/26/19 2:28 PM, PETSc checkBuilds via petsc-checkbuilds wrote:
Dear PETSc developer,
This email contains listings of contributions attributed to you by
`git blame` that caused compiler errors or warnings in PETSc automated
On 4/25/19 6:53 PM, Jed Brown wrote:
Karl Rupp via petsc-dev writes:
With some effort we can certainly address 1.) and to some extent 3.),
probably 4.) as well, but I don't know how to solve 2.) and 5.) with
Jenkins. Given that a significant effort is required for 1.), 3.) and
4.) anyway
Dear PETSc developers,
the current Jenkins server went live last summer. Since then, the
stability of master and next has indeed improved. Who would have thought
three years ago that `next` is almost as stable as `master`?
However, over the weeks and months some weaknesses of our current
Hi Matt,
(...)
His slides have more,
"
PETSc is a widely used library for large sparse iterative solves.
Excellent and comprehensive library of solvers
It is the basis of a significant number of home-made
simulation codes
It is notoriously
Hi Richard,
the check for the GNU compilers is mostly a historic relic. We haven't
done any systematic tests with other compilers, so that test has just
remained in place.
It would certainly be good if you could update the check to also work
well with the default environment on Summit.
arch 22 [i.e
anything that would be acceptable in our maint work-flow shouldn't
be frozen]
- And we should be able to drop troublesome PRs if they are blocking
the release.
full ack :-)
Best regards,
Karli
Satish
On Tue, 5 Mar 2019, Karl Rupp via petsc-dev wrote:
Dear PETSc develo
Dear PETSc developers,
let me suggest Friday, March 22, as the cut-off-date for new Pull
Requests for the upcoming release. This allows for 7 days to iron out
any remaining glitches. (It only took us a few days to release after the
cut-off date last September, so this should be fine)
Also,
Hi,
just for information: we've seen some issues with Thrust in recent CUDA
versions (mostly compilation issues). I don't know whether this is the
cause of this particular error, though.
Best regards,
Karli
On 2/23/19 6:00 AM, Smith, Barry F. via petsc-users wrote:
I get this in the
I have not quickly found how is that "VTK ordering" defined but I
hopefully it's a well-defined unambiguous cell-local numbering. I will
try to find it out soon and get back to you.
Hope this helps:
https://www.vtk.org/wp-content/uploads/2015/04/file-formats.pdf
(page 9)
Best regards,
gt; wrote:
On Wed, Oct 31, 2018 at 5:05 AM Karl Rupp mailto:r...@iue.tuwien.ac.at>> wrote:
Hi Mark,
please comment or remove lines 83 and 84 in
config/BuildSystem/config/packages/cuda.py
Is there a compiler newer than GCC 4.3 available?
Hi Mark,
please comment or remove lines 83 and 84 in
config/BuildSystem/config/packages/cuda.py
Is there a compiler newer than GCC 4.3 available?
Best regards,
Karli
On 10/31/18 8:15 AM, Mark Adams via petsc-dev wrote:
After loading a cuda module ...
On Wed, Oct 31, 2018 at 2:58 AM Mark
Dear PETSc folks,
after broad approval, the pull request for the Code of Conduct has been
merged to master and is now active.
Best regards,
Karli
On 10/23/18 12:52 PM, Karl Rupp wrote:
Dear PETSc folks,
I ask all members of the PETSc team to review the following proposal for
adopting
Dear PETSc folks,
I ask all members of the PETSc team to review the following proposal for
adopting a code of conduct:
https://bitbucket.org/petsc/petsc/pull-requests/1196/code-of-conduct-adopt-contributor-covenant/diff
If you have questions, concerns, etc., please reply to this email
Hi Manuel,
we do not handle more than one GPU per MPI process. If we did, we would
have to duplicate all the existing MPI communication pattern for the
multiple GPUs on a single process. This is a significant maintenance
burden, yet it would not provide any significant performance benefits.
2018, Karl Rupp wrote:
Dear PETSc developers,
please open any outstanding pull requests for the upcoming PETSc 3.10 release
in the next few hours. After that, please do not merge anything to `next` or
`master` unless it is integration work for existing open PRs.
You can open up new pull requ
Hey,
In C we have macros MPIU_INT that we use to indicate that the integer
argument to the MPI call is 64 bit when 64 bit indices are used and 32 bit
otherwise allowing users to write portable code that can just be reconfigured
for 32 or 64 bit integers. I see we do not provide such a
,
Karli
On 08/30/2018 06:09 PM, Karl Rupp wrote:
Dear PETSc developers,
this is a gentle reminder for the cutoff-date on September 4.
Best regards,
Karli
On 07/27/2018 02:41 AM, Karl Rupp wrote:
Dear PETSc developers,
in order to ensure a PETSc 3.10 release no later than by the end
Hi Fabian,
as the error message says, this looks a lot like the GPU running out of
memory. Even though the GTX 560 reports 1 GB of GPU RAM, some of that is
consumed by the OS. Your matrices of size 7024 require about
7*7*sizeof(double) = 392 MB of memory each; this is too much (you can
I thought newer NVIDIA cards had integrated circuits, not using PCI
anymore, still the same case?
NVLink is much faster than PCI, but it's still a bus that is slower (and
higher latency) than aggregate DDR, though there are open questions
about how much message packing (for example) should take
Dear PETSc developers,
this is a gentle reminder for the cutoff-date on September 4.
Best regards,
Karli
On 07/27/2018 02:41 AM, Karl Rupp wrote:
Dear PETSc developers,
in order to ensure a PETSc 3.10 release no later than by the end of
September (possibly earlier), we agreed on September
Hi Manuel,
as Barry said, it is hard for us to provide any help without having a
more complete picture of what is going on.
The error you report seems to come from the AMG preconditioner in
ViennaCL. This can have many origins. Do other preconditioners run
without error? Can you run in a
Hi Youjun,
PetscFortranComplex has been merged to our `master` development branch
and is also available via the `maint` branch. It will be part of the
next 3.9.x point release.
Best regards,
Karli
On 08/04/2018 04:43 PM, Youjun Hu wrote:
Thank Satish.
PetscFortranComplex works.
I just
Hi Santiago,
I am trying to install petsc with the option
--with-precision=__float128. The ./configure goes fine, as well as the
make all stage. However, the make check step to test the libraries fails
with the following error:
/usr/bin/ld:
Hi Fazlul,
Pass the command line option -log_view to get a detailed breakdown of
time spent in various code sections. For details, consult the manual.
Best regards,
Karli
On 07/27/2018 01:47 PM, Fazlul Huq wrote:
Thanks Matthew.
By the way, how can I find the execution time of the
Dear PETSc developers,
in order to ensure a PETSc 3.10 release no later than by the end of
September (possibly earlier), we agreed on September 4, 2018, as the
cut-off date for new features. Please make sure that a pull request has
been opened on Bitbucket by this time. This is the preferred
Hi Pierre,
I’m using GAMG on a shifted Laplacian with these options:
-st_fieldsplit_pressure_ksp_type preonly
-st_fieldsplit_pressure_pc_composite_type additive
-st_fieldsplit_pressure_pc_type composite
-st_fieldsplit_pressure_sub_0_ksp_pc_type jacobi
-st_fieldsplit_pressure_sub_0_pc_type ksp
assed", even though failed tests are reported (timeouts).
2018-07-20 3:35 GMT+02:00 Karl Rupp <mailto:r...@iue.tuwien.ac.at>>:
Hi all,
we now have a first step towards full continuous integration via
Jenkins completed. Thus, every new pull request that is (re-)based
on a
Hi all,
we now have a first step towards full continuous integration via Jenkins
completed. Thus, every new pull request that is (re-)based on a commit
in master not older than today will be automatically tested with a
subset of common tests that are intended to expose the most frequent
Hi Fande,
It looks like SNESTestJacobian() in master is more reliable than that in
maint. Especially, the petsc options name is changed. For example, from
"-snes_test_jacobian_display" to "-snes_test_jacobian_view".
There are some MOOSE tests that fail with the maint version of
Hi,
we're starting to explore (with Andreas cc'd) residual assembly on
GPUs. The question naturally arises: how to do GlobalToLocal and
LocalToGlobal.
I have:
A PetscSF describing the communication pattern.
A Vec holding the data to communicate. This will have an up-to-date
device
Hi all,
(...)Since it looks like MPI endpoints are going to be a long time (or
possibly forever) in coming, I think we need (a) stopgap plan(s) to
support this crappy MPI + OpenMP model in the meantime. One possible
approach is to do what Mark is trying with to do with MKL: Use a third
1 - 100 of 933 matches
Mail list logo