On Tuesday 13 January 2009 03:37, Peter Renzland wrote:
> OK.  Attaching the config_g

I see CONFIG_CROSS_COMPILER_PREFIX="mipsel-uclibc-"

Can you build native i386 or x86_64 busybox with the same .config
(just set CONFIG_CROSS_COMPILER_PREFIX="")?
Does it still exhibit this?

Which libc do you use on your embedded device?



> Shell is ash:
> 
> BusyBox v1.12.2 (2008-12-05 06:20:11 PST) built-in shell (ash)
> also
> BusyBox v1.12.3 (2008-12-14 02:54:58 PST) built-in shell (ash)


# ### it says ash
# ps
  PID USER       VSZ STAT COMMAND
...
21313 root      1984 S    -sh 
21315 root      1960 R    ps 
# ### but ps says sh

So what? your /bin/sh is a symlink to ash, thus /bin/sh is ash.
I see no problem here.


> # ping ac.ca
PING ac.ca (69.90.47.27): 56 data bytes
64 bytes from 69.90.47.27: seq=0 ttl=43 time=132.217 ms
64 bytes from 69.90.47.27: seq=1 ttl=43 time=132.336 ms
# 
--- ac.ca ping statistics ---
2 packets transmitted, 2 packets received, 0% packet loss
round-trip min/avg/max = 132.217/132.276/132.336 ms

# ### SIGINT ^C
# ps
# ### vampire shell eats my command ...
# ps
  PID USER       VSZ STAT COMMAND
    1 root      1720 S    init noinitrd 
    2 root         0 SW   [keventd]



Mine doesn't:

# ping ac.ca
PING ac.ca (69.90.47.27): 56 data bytes
64 bytes from 69.90.47.27: seq=0 ttl=47 time=181.496 ms
64 bytes from 69.90.47.27: seq=1 ttl=47 time=180.338 ms
^C
--- ac.ca ping statistics ---
3 packets transmitted, 2 packets received, 33% packet loss
round-trip min/avg/max = 180.338/180.917/181.496 ms

# ps
PID   USER     TIME   COMMAND
    1 root       0:00 /bin/sh /init
    2 root       0:00 [kthreadd]
    3 root       0:00 [migration/0]



Having said that, these is a few differences in your and my setup:

* I use native build against glibc
* I use svn, which have a few fixes in ash.

I will mail you latest terball separately. Please try it and let me know.
--
vda
_______________________________________________
busybox mailing list
[email protected]
http://lists.busybox.net/mailman/listinfo/busybox

Reply via email to