On Fri, May 03, 2019 at 05:51:54PM +0200, Phil Sutter wrote:
> Hi,
>
> On Wed, May 01, 2019 at 12:35:00PM -0400, Eric Garver wrote:
> > When calling ffi functions we need to convert from python strings to
> > utf-8. Then convert back for any output we receive.
>
> So the problem is passing utf-8 encoded strings as command?
In python3 strings are unicode. But we need "bytes" when calling the
ctypes function since it's imported with "c_char_p". This is what
encode() is doing for us.
https://docs.python.org/3/library/ctypes.html#fundamental-data-types
In python2 strings are a sequence of bytes already. I'll have to v2 to
if we care about python2 support.
>
> [...]
> > - rc = self.nft_run_cmd_from_buffer(self.__ctx, cmdline)
> > - output = self.nft_ctx_get_output_buffer(self.__ctx)
> > - error = self.nft_ctx_get_error_buffer(self.__ctx)
> > + rc = self.nft_run_cmd_from_buffer(self.__ctx,
> > cmdline.encode("utf-8"))
> > + output = self.nft_ctx_get_output_buffer(self.__ctx).decode("utf-8")
> > + error = self.nft_ctx_get_error_buffer(self.__ctx).decode("utf-8")
>
> Should the encoding be made configurable? I see encode() and decode()
> parameters are optional, but as soon as I call them with a string
> containing umlauts I get errors. So not sure if that would be an
> alternative.
I don't think so. Since we're calling system level stuff (nftables,
kernel) I think utf-8 is what we want.
Encoding with utf-8 does the right thing:
python3:
>>> "ö".encode("utf-8")
>>> b'\xc3\xb6'
python2:
>>> u"ö".encode("utf-8")
>>> '\xc3\xb6'