>From folloing logs, we can see there are multiple incoming requests in the 
same second from the same host:port

[2022-08-08 21:30:18.707037] call_data 0x55555b09b810 from host 
ipv4:192.168.0.42:50576                                                     
     
 [2022-08-08 21:30:18.724226] call_data 0x55555b098ca0 from host 
ipv4:192.168.0.42:50576                                                     
      
[2022-08-08 21:30:18.762191] call_data 0x55555b09f4a0 from host 
ipv4:192.168.0.42:38126                                                     
      
[2022-08-08 21:30:18.762573] call_data 0x55555b1606a0 from host 
ipv4:192.168.0.42:38126                                                     
      
[2022-08-08 21:30:18.762507] call_data 0x55555b0a0600 from host 
ipv4:192.168.0.42:38126
[2022-08-08 21:30:18.762365] call_data 0x55555b187aa0 from host 
ipv4:192.168.0.42:38126                                                     
      
[2022-08-08 21:30:18.767045] call_data 0x55555b09a6b0 from host 
ipv4:192.168.0.42:50576                                                     
     
 [2022-08-08 21:30:18.768973] call_data 0x55555b0dc0e0 from host 
ipv4:192.168.0.56:56830                                                     
      
[2022-08-08 21:30:18.769633] call_data 0x55555b09af20 from host 
ipv4:192.168.0.56:56830                                                     
      
[2022-08-08 21:30:18.771011] call_data 0x55555b09c080 from host 
ipv4:192.168.0.56:56830                                                     
     
 [2022-08-08 21:30:18.786424] call_data 0x55555b097290 from host 
ipv4:192.168.0.42:50576

If I lock the mutex before calling "Finish", the issue would be gone, and 
the side effect is performance downgrade.

If I manually set channel argument GRPC_ARG_MAX_CONCURRENT_STREAMS for both 
client and side, the issue is gone as well.

Does anyone know the reason?


On Monday, August 8, 2022 at 1:36:08 PM UTC+8 Youbing Li wrote:

> Hello,
>
>  My service has multiple methods.
>
> Async C++ API is using and those methods shared the same Completion Queue 
> in the server side.
>
> In each calldata, there is an object 
>     grpc::ServerAsyncResponseWriter<Reply> responder_{nullptr}
>
> In the application layer, each calldata is used by one thread.
> If multiple threads simultaneously call responder_->Finish, the server 
> will abort.
>
> It sames like multiple operations on same completion queue is not 
> supported.
>
> GDB stack: (Note: multiple threads call *Finish* in all stacks)
>
> 0  __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:50
>
> #1  0x00007ffff76c9859 in __GI_abort () at abort.c:79
>
> #2  0x0000555559003b21 in 
> absl::lts_20211102::raw_logging_internal::(anonymous 
> namespace)::RawLogVA(absl::lts_20211102::LogSeverity, const char *, int, 
> const char *, typedef __va_list_tag __va_list_tag *) 
> (severity=absl::lts_20211102::LogSeverity::kFatal, file=0x5555596fdcca 
> "mutex.cc", line=1922, format=0x5555596fdb58 "Check %s failed: %s", 
> ap=0x7ffc5a9eaef0) at 
> /data/src/open-src/vcpkg/buildtrees/abseil/src/cae233d0bf-0b16e4be7c.clean/absl/base/internal/raw_logging.cc:182
>
> #3  0x0000555559003c98 in absl::lts_20211102::raw_logging_internal::RawLog 
> (severity=absl::lts_20211102::LogSeverity::kFatal, file=0x5555596fdcca 
> "mutex.cc", line=1922, format=0x5555596fdb58 "Check %s failed: %s") at 
> /data/src/open-src/vcpkg/buildtrees/abseil/src/cae233d0bf-0b16e4be7c.clean/absl/base/internal/raw_logging.cc:216
>
> #4  0x0000555558f9c962 in absl::lts_20211102::Mutex::LockSlowLoop 
> (this=0x55555af72d28, waitp=0x7ffc5a9eb1f0, flags=0) at 
> /data/src/open-src/vcpkg/buildtrees/abseil/src/cae233d0bf-0b16e4be7c.clean/absl/synchronization/mutex.cc:1922
>
> #5  0x0000555558f9c679 in absl::lts_20211102::Mutex::LockSlowWithDeadline 
> (this=0x55555af72d28, how=0x5555596fde60 <absl::lts_20211102::kExclusiveS>, 
> cond=0x0, t=..., flags=0) at 
> /data/src/open-src/vcpkg/buildtrees/abseil/src/cae233d0bf-0b16e4be7c.clean/absl/synchronization/mutex.cc:1875
>
> #6  0x0000555558f9c33b in absl::lts_20211102::Mutex::LockSlow 
> (this=0x55555af72d28, how=0x5555596fde60 <absl::lts_20211102::kExclusiveS>, 
> cond=0x0, flags=0) at 
> /data/src/open-src/vcpkg/buildtrees/abseil/src/cae233d0bf-0b16e4be7c.clean/absl/synchronization/mutex.cc:1768
>
> #7  0x0000555558f9b309 in absl::lts_20211102::Mutex::Lock 
> (this=0x55555af72d28) at 
> /data/src/open-src/vcpkg/buildtrees/abseil/src/cae233d0bf-0b16e4be7c.clean/absl/synchronization/mutex.cc:1481
>
> #8  0x0000555558ed7038 in gpr_mu_lock (mu=0x55555af72d28) at 
> /data/src/open-src/vcpkg/buildtrees/grpc/src/85a295989c-6cf7bf442d.clean/src/core/lib/gpr/sync_abseil.cc:56
>
> #9  0x0000555558c001c8 in grpc_cq_begin_op (cq=0x55555af72c20, 
> tag=0x55555b062210) at 
> /data/src/open-src/vcpkg/buildtrees/grpc/src/85a295989c-6cf7bf442d.clean/src/core/lib/surface/completion_queue.cc:666
>
> #10 0x0000555558bf0d49 in call_start_batch (call=0x7ffff0016130, 
> ops=0x7ffc5a9eb590, nops=3, notify_tag=0x55555b062210, 
> is_notify_tag_closure=0) at 
> /data/src/open-src/vcpkg/buildtrees/grpc/src/85a295989c-6cf7bf442d.clean/src/core/lib/surface/call.cc:1702
>
> #11 0x0000555558bf10c1 in grpc_call_start_batch (call=0x7ffff0016130, 
> ops=0x7ffc5a9eb590, nops=3, tag=0x55555b062210, reserved=0x0) at 
> /data/src/open-src/vcpkg/buildtrees/grpc/src/85a295989c-6cf7bf442d.clean/src/core/lib/surface/call.cc:1761
>
> #12 0x000055555642e799 in grpc::CoreCodegen::grpc_call_start_batch 
> (this=0x55555ae90ad0, call=0x7ffff0016130, ops=0x7ffc5a9eb590, nops=3, 
> tag=0x55555b062210, reserved=0x0) at 
> /data/src/open-src/vcpkg/buildtrees/grpc/src/85a295989c-6cf7bf442d.clean/src/cpp/common/core_codegen.cc:113
>
> #13 0x0000555555c4ed5e in 
> grpc::internal::CallOpSet<grpc::internal::CallOpSendInitialMetadata, 
> grpc::internal::CallOpSendMessage, grpc::internal::CallOpServerSendStatus, 
> grpc::internal::CallNoOp<4>, grpc::internal::CallNoOp<5>, 
> grpc::internal::CallNoOp<6> >::ContinueFillOpsAfterInterception 
> (this=0x55555b062210) at 
> /data/src/fabarta/fabarta-compute/build/vcpkg_installed/x64-linux/include/grpcpp/impl/codegen/call_op_set.h:971
>
> #14 0x0000555555c4eb05 in 
> grpc::internal::CallOpSet<grpc::internal::CallOpSendInitialMetadata, 
> grpc::internal::CallOpSendMessage, grpc::internal::CallOpServerSendStatus, 
> grpc::internal::CallNoOp<4>, grpc::internal::CallNoOp<5>, 
> grpc::internal::CallNoOp<6> >::FillOps (this=0x55555b062210, 
> call=0x55555b062078) at 
> /data/src/fabarta/fabarta-compute/build/vcpkg_installed/x64-linux/include/grpcpp/impl/codegen/call_op_set.h:901
>
> #15 0x0000555556445e46 in grpc::Server::PerformOpsOnCall 
> (this=0x55555af57160, ops=0x55555b062210, call=0x55555b062078) at 
> /data/src/open-src/vcpkg/buildtrees/grpc/src/85a295989c-6cf7bf442d.clean/src/cpp/server/server_cc.cc:1305
>
> #16 0x0000555555c2198d in grpc::internal::Call::PerformOps 
> (this=0x55555b062078, ops=0x55555b062210) at 
> /data/src/fabarta/fabarta-compute/build/vcpkg_installed/x64-linux/include/grpcpp/impl/codegen/call.h:68
>
> #17 0x0000555555c4f9c9 in 
> grpc::ServerAsyncResponseWriter<fabarta::compute::ReplyMsg>::Finish 
> (this=0x55555b062070, msg=..., status=..., tag=0x55555b061bc0) at 
> /data/src/fabarta/fabarta-compute/build/vcpkg_installed/x64-linux/include/grpcpp/impl/codegen/async_unary_call.h:355
>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"grpc.io" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to grpc-io+unsubscr...@googlegroups.com.
To view this discussion on the web visit 
https://groups.google.com/d/msgid/grpc-io/329f8f5f-3009-4d54-be24-3ce904da88cfn%40googlegroups.com.

Reply via email to