Date   

Re: how to reuse generated library in a nativesdk recipe #sdk #systemd

Mikko Rapeli
 

Hi,

On Wed, Feb 19, 2020 at 01:37:19AM -0800, Armando Hernandez wrote:
Hello,

I have a recipe that builds a library. The recipe specifies an additional package "${PN}-systemd" along with other systemd related variables and finally it instructs that the package should be built with "-DWITH_SYSTEMD=ON" being passed to cmake. So far so good. But, I extended this recipe to nativesdk because I need this library on it. When trying to build the corresponding nativesdk package, the build fails at the configuration step (i.e. "do_configure") claiming it cannot find the package systemd.

Is there a way I can install the -already-generated libraries into my SDK (potentially via the corresponding nativesdk recipe) without having to rebuild the package? Or do I need to somehow include such systemd package in my sdk (which I don't think I need at all)?

Any hints and pointers as to were to look at are very well appreciated.
Thanks.
Make the systemd dependency for target only, e.g. DEPENDS_class-target += "systemd"
etc.

There may be relevant use cases to build some of systemd components or tools
to native or nativesdk targets too. In that case add BBCLASSEXTEND += "nativesdk" etc
in a bbappend to systemd.

Hope this helps,

-Mikko


Re: Opentk Support

Ross Burton
 

On Wed, 19 Feb 2020 at 21:45, Sheraz Ali <sheraz.ali@iwavesystems.com> wrote:
Does anyone know how to enable opentk in yocto ( i.e is it available )
https://layers.openembedded.org/layerindex/branch/master/recipes/?q=opentk
says that there are not any known recipes, so you'll have to write one
yourself.

Ross


Re: [OE-core] oe-core recipe for defining directories in /

Quentin Schulz
 

Hi JH,

On Wed, Feb 19, 2020 at 09:12:08PM +1100, JH wrote:
Hi,

Which recipe defines all directories in "/"? I need to make a bbapand
to add directories to /.
None. Or all of them, depends on how one sees it.

You just create a directory in do_install of a recipe.

You then make sure this directory is part of a package by checking it's
in one of the recipe's generated packages's FILES_<PACKAGE>.

Quentin


Zeus failed DHCP

JH
 

Hi,

My build connman on Thud works on WiFi, but Zeus does not work, the
connman could not get WiFi DHCP response, it puts a local IP address
169.254.24.188 to my WiFi interface. Has anyone found that the problem
in Zeus or it just me may be miss some packages or configuration?

What are packages could cause DHCP not working?

Here is Thud working log:

# systemctl status connman -l
��● connman.service - Connection service
Loaded: loaded
(8;;file://solar/lib/systemd/system/connman.service/lib/systemd/system/connman.service8;;;
enabl)
Active: active (running) since Thu 2020-02-13 03:18:51 UTC; 6 days ago
Main PID: 131 (connmand)
CGroup: /system.slice/connman.service
��└��─131 /usr/sbin/connmand -n

Feb 18 22:37:16 solar connmand[131]: rp_filter set to 2 (loose mode
routing), old value was 1
Feb 18 22:37:16 solar connmand[131]: mlan0 {add} address
192.168.0.100/24 label mlan0 family 2
Feb 18 22:37:16 solar connmand[131]: mlan0 {add} route 192.168.0.0 gw
0.0.0.0 scope 253 <LINK>
Feb 18 22:37:16 solar connmand[131]: mlan0 {add} route 192.168.0.1 gw
0.0.0.0 scope 253 <LINK>
Feb 18 22:37:16 solar connmand[131]: mlan0 {add} route 212.227.81.55
gw 192.168.0.1 scope 0 <UNIVERSE>
Feb 18 22:37:17 solar connmand[131]: mlan0 {del} route 212.227.81.55
gw 192.168.0.1 scope 0 <UNIVERSE>
Feb 18 22:37:17 solar connmand[131]: wwan0 {del} route 0.0.0.0 gw
10.114.57.126 scope 0 <UNIVERSE>
Feb 18 22:37:17 solar connmand[131]: mlan0 {add} route 0.0.0.0 gw
192.168.0.1 scope 0 <UNIVERSE>
Feb 18 22:37:17 solar connmand[131]: mlan0 {add} route 212.227.81.55
gw 192.168.0.1 scope 0 <UNIVERSE>
Feb 18 22:37:17 solar connmand[131]: mlan0 {del} route 212.227.81.55
gw 192.168.0.1 scope 0 <UNIVERSE>

# systemctl status wpa_supplicant -l
��● wpa_supplicant.service - WPA supplicant
Loaded: loaded
(8;;file://solar/lib/systemd/system/wpa_supplicant.service/lib/systemd/system/wpa_supplicant.service8;;;
disabled; vendor preset: enabled)
Active: active (running) since Thu 2020-02-13 03:18:53 UTC; 6 days ago
Main PID: 503 (wpa_supplicant)
CGroup: /system.slice/wpa_supplicant.service
��└��─503 /usr/sbin/wpa_supplicant -u

Feb 19 07:37:29 solar wpa_supplicant[503]: mlan0: WPA: Group rekeying
completed with 34:08:04:12:b1:a2 [GTK=TKIP]
Feb 19 07:47:29 solar wpa_supplicant[503]: mlan0: WPA: Group rekeying
completed with 34:08:04:12:b1:a2 [GTK=TKIP]
Feb 19 07:57:29 solar wpa_supplicant[503]: mlan0: WPA: Group rekeying
completed with 34:08:04:12:b1:a2 [GTK=TKIP]
Feb 19 08:07:29 solar wpa_supplicant[503]: mlan0: WPA: Group rekeying
completed with 34:08:04:12:b1:a2 [GTK=TKIP]
Feb 19 08:17:29 solar wpa_supplicant[503]: mlan0: WPA: Group rekeying
completed with 34:08:04:12:b1:a2 [GTK=TKIP]
Feb 19 08:19:01 solar wpa_supplicant[503]: mlan0:
CTRL-EVENT-SIGNAL-CHANGE above=0 signal=-85 noise=-97 txrate=1000
Feb 19 08:27:29 solar wpa_supplicant[503]: mlan0: WPA: Group rekeying
completed with 34:08:04:12:b1:a2 [GTK=TKIP]
Feb 19 08:37:29 solar wpa_supplicant[503]: mlan0: WPA: Group rekeying
completed with 34:08:04:12:b1:a2 [GTK=TKIP]
Feb 19 08:47:29 solar wpa_supplicant[503]: mlan0: WPA: Group rekeying
completed with 34:08:04:12:b1:a2 [GTK=TKIP]
Feb 19 08:57:29 solar wpa_supplicant[503]: mlan0: WPA: Group rekeying
completed with 34:08:04:12:b1:a2 [GTK=TKIP]

Here is Zenu build did not work:

# systemctl status connman -l
* connman.service - Connection service
Loaded: loaded
(8;;file://solar/lib/systemd/system/connman.service/lib/systemd/system/connman.service8;;;
enabled; vendor preset: enabled)
Active: active (running) since Tue 2020-02-18 00:47:43 UTC; 1 day 9h ago
Main PID: 184 (connmand)
CGroup: /system.slice/connman.service
`-184 /usr/sbin/connmand -n

Feb 19 10:27:33 solar connmand[184]: mlan0 {newlink} index 3 operstate
5 <DORMANT>
Feb 19 10:27:33 solar connmand[184]: mlan0 {add} route ff00:: gw ::
scope 0 <UNIVERSE>
Feb 19 10:27:33 solar connmand[184]: mlan0 {add} route fe80:: gw ::
scope 0 <UNIVERSE>
Feb 19 10:27:33 solar connmand[184]: mlan0 {RX} 10 packets 1650 bytes
Feb 19 10:27:33 solar connmand[184]: mlan0 {TX} 256 packets 96668 bytes
Feb 19 10:27:33 solar connmand[184]: mlan0 {update} flags 102467
<UP,RUNNING,LOWER_UP>
Feb 19 10:27:33 solar connmand[184]: mlan0 {newlink} index 3 address
D4:CA:6E:9A:7E:29 mtu 1500
Feb 19 10:27:33 solar connmand[184]: mlan0 {newlink} index 3 operstate 6 <UP>
Feb 19 10:28:13 solar connmand[184]: mlan0 {add} address
169.254.241.106/16 label mlan0 family 2
Feb 19 10:28:14 solar connmand[184]: mlan0 {add} route 169.254.0.0 gw
0.0.0.0 scope 253 <LINK>


# systemctl status wpa_supplicant -l
* wpa_supplicant.service - WPA supplicant
Loaded: loaded
(8;;file://solar/lib/systemd/system/wpa_supplicant.service/lib/systemd/system/wpa_supplicant.service8;;;
disabled; vendor preset: disabled)
Active: active (running) since Tue 2020-02-18 00:47:48 UTC; 1 day 9h ago
Main PID: 263 (wpa_supplicant)
CGroup: /system.slice/wpa_supplicant.service
`-263 /usr/sbin/wpa_supplicant -u

Feb 19 09:57:33 solar wpa_supplicant[263]: mlan0:
CTRL-EVENT-SIGNAL-CHANGE above=0 signal=-47 noise=-92 txrate=65000
Feb 19 10:05:36 solar wpa_supplicant[263]: mlan0:
CTRL-EVENT-SIGNAL-CHANGE above=0 signal=-52 noise=-92 txrate=72200
Feb 19 10:07:32 solar wpa_supplicant[263]: mlan0:
CTRL-EVENT-DISCONNECTED bssid=34:08:04:12:b1:a2 reason=2
Feb 19 10:07:32 solar wpa_supplicant[263]: dbus:
wpa_dbus_property_changed: no property SessionLength in object
/fi/w1/wpa_supplicant1/Interfaces/0
Feb 19 10:07:33 solar wpa_supplicant[263]: mlan0: Trying to associate
with 34:08:04:12:b1:a2 (SSID='Jupiter' freq=2437 MHz)
Feb 19 10:07:33 solar wpa_supplicant[263]: mlan0: Associated with
34:08:04:12:b1:a2
Feb 19 10:07:33 solar wpa_supplicant[263]: mlan0:
CTRL-EVENT-SUBNET-STATUS-UPDATE status=0
Feb 19 10:07:33 solar wpa_supplicant[263]: mlan0: WPA: Key negotiation
completed with 34:08:04:12:b1:a2 [PTK=CCMP GTK=TKIP]
Feb 19 10:07:33 solar wpa_supplicant[263]: mlan0: CTRL-EVENT-CONNECTED
- Connection to 34:08:04:12:b1:a2 completed [id=0 id_str=]
Feb 19 10:07:33 solar wpa_supplicant[263]: mlan0:
CTRL-EVENT-SIGNAL-CHANGE above=0 signal=-51 noise=-92 txrate=65000

Thank


Re: [OE-core] [yocto] Change RO rootfs failed RF Kill Switch Status and Failed to start Run pending postinsts

Mikko Rapeli
 

On Tue, Feb 18, 2020 at 08:43:01PM +1100, JH wrote:
Hi Mikko,

On 2/18/20, Mikko.Rapeli@bmw.de <Mikko.Rapeli@bmw.de> wrote:
I think you may be missing volatile-binds package and service from your
image.
See poky/meta/recipes-core/volatile-binds/volatile-binds.bb
I got the source in my build system, it is zeus:

oe-core/meta/recipes-core/volatile-binds/volatile-binds.bb

./all-oe-linux/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/packages-split/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/sysroot-destdir/sysroot-providers/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata-pdata-input/runtime/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata-pdata-input/runtime-reverse/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata-pdata-input/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata/runtime/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata/runtime-reverse/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/license-destdir/volatile-binds

Are there correct?

It may be missing /var/log but with systemd there should not be needs to
write
to that location after image builds...
The volatile did not have the log, so /var/log -> volatile/log was an
invalid link, should I manually create it?

Thanks Mikko,
Well I have zeus and am using read-only rootfs with volatile binds and
I did not need anything extra. I would dig into this /var/log thing
and patch it away. I use systemd journal so no need for syslogs.

Cheers,

-Mikko


Re: [opkg-devel] [opkg-utils PATCH] Makefile: add opkg-feed to UTILS

Alejandro del Castillo <alejandro.delcastillo@...>
 

LGTM, merged!

On 2/17/20 5:57 PM, Alex Stewart wrote:
* Add the opkg-feed script to UTILS so that it is installed with a `make
install`.
* Clean up the UTILS variable declaration to be a little more diffable.
Signed-off-by: Alex Stewart <alex.stewart@ni.com>
---
Makefile | 17 ++++++++++++++---
1 file changed, 14 insertions(+), 3 deletions(-)
diff --git a/Makefile b/Makefile
index 817a8c1..4049654 100644
--- a/Makefile
+++ b/Makefile
@@ -1,6 +1,17 @@
-UTILS = opkg-build opkg-unbuild opkg-make-index opkg.py opkg-list-fields \
- arfile.py opkg-buildpackage opkg-diff opkg-extract-file opkg-show-deps \
- opkg-compare-indexes update-alternatives
+UTILS = \
+ arfile.py \
+ opkg-build \
+ opkg-buildpackage \
+ opkg-compare-indexes \
+ opkg-diff \
+ opkg-extract-file \
+ opkg-feed \
+ opkg-list-fields \
+ opkg-make-index \
+ opkg-show-deps \
+ opkg-unbuild \
+ opkg.py \
+ update-alternatives
MANPAGES = opkg-build.1
--
Cheers,

Alejandro


Re: Change RO rootfs failed RF Kill Switch Status and Failed to start Run pending postinsts

JH
 

It also seems mwifiex_sdio tried to write to RO rootfs and failed and
triggled RF Killm, does mwifiex_sdio needs some system directories for
RW?

[ 26.636845] mwifiex_sdio mmc0:0001:1: mwifiex_process_cmdresp: cmd 0x242 fain
Starting Load/Save RF Kill Switch Status...
[ 26.852990] mwifiex_sdio mmc0:0001:1: info: MWIFIEX VERSION: mwifiex 1.0 (14
[ 26.861518] mwifiex_sdio mmc0:0001:1: driver_version = mwifiex 1.0 (14.68.36
[FAILED] Failed to start Load/Save RF Kill Switch Status.
See 'systemctl status systemd-rfkill.service' for details.
Starting Load/Save RF Kill Switch Status...
[FAILED] Failed to start Load/Save RF Kill Switch Status.
See 'systemctl status systemd-rfkill.service' for details.
Starting Load/Save RF Kill Switch Status...
[FAILED] Failed to start Load/Save RF Kill Switch Status.
See 'systemctl status systemd-rfkill.service' for details.
Starting Load/Save RF Kill Switch Status...

On 2/18/20, JH <jupiter.hce@gmail.com> wrote:
Hi,

Apologize for the cross posting.

I am running kernel 4.19.75 on iMX6 customized device with WiFi and 4G
LTE, it was running well in an RW rootfs. After I have just changed
rootfs to RO UBIFS partition, it failed RF Kill and postinsts I
suspect both try write to the RO and failed, any advice how to fix it?
Despite it failed RF Kill and postinsts, it was still working.

[ 6.097762] UBIFS (ubi0:2): UBIFS: mounted UBI device 0, volume 2,
name "rootfs-volume", R/O mode
..............
[ 6.151932] VFS: Mounted root (ubifs filesystem) readonly on device
0:13.
.................
[ OK ] Listening on Load/Save RF Kill Switch Status /dev/rfkill Watch.
Starting Load/Save RF Kill Switch Status...
[FAILED] Failed to start Load/Save RF Kill Switch Status.
See 'systemctl status systemd-rfkill.service' for details.

[FAILED] Failed to start Run pending postinsts.
See 'systemctl status run-postinsts.service' for details.
...............
root#

Thank you.

Kind regards,

- jh


Re: Change RO rootfs failed RF Kill Switch Status and Failed to start Run pending postinsts

Marek Belisko
 

Hi,

On Tue, Feb 18, 2020 at 2:00 AM JH <jupiter.hce@gmail.com> wrote:

Hi,

Apologize for the cross posting.

I am running kernel 4.19.75 on iMX6 customized device with WiFi and 4G
LTE, it was running well in an RW rootfs. After I have just changed
rootfs to RO UBIFS partition, it failed RF Kill and postinsts I
suspect both try write to the RO and failed, any advice how to fix it?
Despite it failed RF Kill and postinsts, it was still working.

[ 6.097762] UBIFS (ubi0:2): UBIFS: mounted UBI device 0, volume 2,
name "rootfs-volume", R/O mode
..............
[ 6.151932] VFS: Mounted root (ubifs filesystem) readonly on device 0:13.
.................
[ OK ] Listening on Load/Save RF Kill Switch Status /dev/rfkill Watch.
Starting Load/Save RF Kill Switch Status...
[FAILED] Failed to start Load/Save RF Kill Switch Status.
See 'systemctl status systemd-rfkill.service' for details.
Can you pls provide output of systemctl status systemd-rfkill
There should be some more info what issue is.

[FAILED] Failed to start Run pending postinsts.
See 'systemctl status run-postinsts.service' for details.
Pls this one also: systemctl status run-postinsts
...............
root#

Thank you.

Kind regards,

- jh
BR,

marek


Re: Change RO rootfs failed RF Kill Switch Status and Failed to start Run pending postinsts

JH
 

Hi Belisko,

Thanks for your resonse.

On 2/18/20, Belisko Marek <marek.belisko@gmail.com> wrote:
Can you pls provide output of systemctl status systemd-rfkill
There should be some more info what issue is.
Failed at step STATE_DIRECTORY spawning /lib/systemd/systemd-rfkill:
Read-only file system, did it try to write something in /lib/systemd?
How should I fix it?


# systemctl status systemd-rfkill -l
* systemd-rfkill.service - Load/Save RF Kill Switch Status
Loaded: loaded
(8;;file://solar/lib/systemd/system/systemd-rfkill.service/lib/systemd/system/systemd-rfkill.service8;;;
static; vendor preset: disabled)
Active: failed (Result: exit-code) since Tue 2020-02-18 00:47:30
UTC; 1min 59s ago
Docs: 8;;man:systemd-rfkill.service(8)man:systemd-rfkill.service(8)8;;
Process: 149 ExecStart=/lib/systemd/systemd-rfkill (code=exited,
status=238/STATE_DIRECTORY)
Main PID: 149 (code=exited, status=238/STATE_DIRECTORY)

Feb 18 00:47:30 solar systemd[1]: Starting Load/Save RF Kill Switch Status...
Feb 18 00:47:30 solar systemd[149]: systemd-rfkill.service: Failed to
set up special execution directory in /var/lib: Read-only file system
Feb 18 00:47:30 solar systemd[149]: systemd-rfkill.service: Failed at
step STATE_DIRECTORY spawning /lib/systemd/systemd-rfkill: Read-only
file system
Feb 18 00:47:30 solar systemd[1]: systemd-rfkill.service: Main process
exited, code=exited, status=238/STATE_DIRECTORY
Feb 18 00:47:30 solar systemd[1]: systemd-rfkill.service: Failed with
result 'exit-code'.
Feb 18 00:47:30 solar systemd[1]: Failed to start Load/Save RF Kill
Switch Status.
Feb 18 00:47:30 solar systemd[1]: systemd-rfkill.service: Start
request repeated too quickly.
Feb 18 00:47:30 solar systemd[1]: systemd-rfkill.service: Failed with
result 'exit-code'.
Feb 18 00:47:30 solar systemd[1]: Failed to start Load/Save RF Kill
Switch Status.

[FAILED] Failed to start Run pending postinsts.
See 'systemctl status run-postinsts.service' for details.
Pls this one also: systemctl status run-postinsts
# systemctl status run-postinsts -l
* run-postinsts.service - Run pending postinsts
Loaded: loaded
(8;;file://solar/lib/systemd/system/run-postinsts.service/lib/systemd/system/run-postinsts.service8;;;
enabled; vendor preset: enabled)
Active: failed (Result: exit-code) since Tue 2020-02-18 00:47:37
UTC; 6min ago
Process: 153 ExecStart=/usr/sbin/run-postinsts (code=exited, status=0/SUCCESS)
Process: 159 ExecStartPost=/bin/systemctl --no-reload disable
run-postinsts.service (code=exited, status=1/FAILURE)
Main PID: 153 (code=exited, status=0/SUCCESS)

Feb 18 00:47:36 solar systemd[1]: Starting Run pending postinsts...
Feb 18 00:47:36 solar run-postinsts[153]: Configuring packages on first boot....
Feb 18 00:47:36 solar run-postinsts[153]: (This may take several
minutes. Please do not power off the machine.)
Feb 18 00:47:36 solar run-postinsts[153]: /usr/sbin/run-postinsts:
eval: line 1: can't create /var/log/postinstall.log: nonexistent
directory
Feb 18 00:47:36 solar run-postinsts[153]: Removing any system startup
links for run-postinsts ...
Feb 18 00:47:37 solar systemctl[159]: Failed to disable unit: File
/etc/systemd/system/sysinit.target.wants/run-postinsts.service:
Read-only file system
Feb 18 00:47:37 solar systemd[1]: run-postinsts.service: Control
process exited, code=exited, status=1/FAILURE
Feb 18 00:47:37 solar systemd[1]: run-postinsts.service: Failed with
result 'exit-code'.
Feb 18 00:47:37 solar systemd[1]: Failed to start Run pending postinsts.

Was the problem to write to /var/log, the /var/volatile does not have a log?

# ls -l /var
drwxr-xr-x 2 1000 1000 160 Feb 18 2020 backups
drwxr-xr-x 5 1000 1000 100 Feb 18 00:47 cache
drwxr-xr-x 9 1000 1000 180 Feb 18 00:47 lib
drwxr-xr-x 3 1000 1000 224 Feb 18 2020 local
lrwxrwxrwx 1 1000 1000 11 Feb 18 2020 lock -> ../run/lock
lrwxrwxrwx 1 1000 1000 12 Feb 18 00:52 log -> volatile/log
lrwxrwxrwx 1 1000 1000 6 Feb 18 2020 run -> ../run
drwxr-xr-x 3 1000 1000 60 Feb 18 2020 spool
lrwxrwxrwx 1 1000 1000 12 Feb 18 2020 tmp -> volatile/tmp
drwxrwxrwt 8 root root 160 Feb 18 00:47 volatile

# ls -l /var/volatile/
drwxr-xr-x 5 1000 1000 100 Feb 18 00:47 cache
drwxr-xr-x 9 1000 1000 180 Feb 18 00:47 lib
drwxr-xr-x 3 1000 1000 60 Feb 18 2020 spool

All system mount is the same as the original RW rootfs, did both write
to none standard RW system mount?

Here is defined system mount in fstab:

proc /proc proc defaults 0 0
devpts /dev/pts devpts mode=0620,gid=5 0 0
tmpfs /run tmpfs
mode=0755,nodev,nosuid,strictatime 0 0
tmpfs /var/volatile tmpfs defaults 0 0


Here is the mount:

# mount
ubi0:rootfs-volume on / type ubifs (ro,relatime,assert=read-only,ubi=0,vol=2)
devtmpfs on /dev type devtmpfs
(rw,relatime,size=84564k,nr_inodes=21141,mode=755)
sysfs on /sys type sysfs (rw,nosuid,nodev,noexec,relatime)
proc on /proc type proc (rw,relatime)
securityfs on /sys/kernel/security type securityfs
(rw,nosuid,nodev,noexec,relatime)
tmpfs on /dev/shm type tmpfs (rw,nosuid,nodev)
devpts on /dev/pts type devpts (rw,relatime,gid=5,mode=620,ptmxmode=000)
tmpfs on /run type tmpfs (rw,nosuid,nodev,mode=755)
tmpfs on /sys/fs/cgroup type tmpfs (ro,nosuid,nodev,noexec,mode=755)
cgroup2 on /sys/fs/cgroup/unified type cgroup2
(rw,nosuid,nodev,noexec,relatime,nsdelegate)
cgroup on /sys/fs/cgroup/systemd type cgroup
(rw,nosuid,nodev,noexec,relatime,xattr,name=systemd)
tmpfs on /etc/machine-id type tmpfs (ro,mode=755)
tmpfs on /tmp type tmpfs (rw,nosuid,nodev)
debugfs on /sys/kernel/debug type debugfs (rw,nosuid,nodev,noexec,relatime)
fusectl on /sys/fs/fuse/connections type fusectl
(rw,nosuid,nodev,noexec,relatime)
tmpfs on /var/volatile type tmpfs (rw,relatime)
ubi0:data-volume on /data type ubifs (rw,noatime,assert=read-only,ubi=0,vol=3)
tmpfs on /var/spool type tmpfs (rw,relatime)
tmpfs on /var/cache type tmpfs (rw,relatime)
tmpfs on /var/lib type tmpfs (rw,relatime)
tracefs on /sys/kernel/debug/tracing type tracefs
(rw,nosuid,nodev,noexec,relatime)

How should I fix it?

Thank you.

Kind regards,

- jh


Re: [OE-core] [yocto] Change RO rootfs failed RF Kill Switch Status and Failed to start Run pending postinsts

Mikko Rapeli
 

(trimming lists to yocto only)

Hi,

I think you may be missing volatile-binds package and service from your image.
See poky/meta/recipes-core/volatile-binds/volatile-binds.bb

It may be missing /var/log but with systemd there should not be needs to write
to that location after image builds...

-Mikko


Debugging gdb built by Yocto

Patrick Doyle <wpdster@...>
 

Does anybody have any tips or tricks for how I might debug the
(cross-canadian) gdb built by Yocto's SDK?

I need to add some printf's to the gdb code to help track down why
something isn't working, but none of my traditional
get-ready-to-debug-this-code techniques are working.

How can I run the gdb that I just built? Note that I am presuming that I can

$ bitbake gdb-cross-canadian-mipsel -ccompile -f

to rebuild gdb after I add a printf or two to it... but I can't figure
out how to run gdb without going through the sdk installation step.

$ bitbake gdb-cross-canadian-mipsel -cdevshell
# ../build-mipsel-poky-linux/gdb/gdb
bash: ../build-mipsel-poky-linux/gdb/gdb: No such file or directory
# file ../build-mipsel-poky-linux/gdb/gdb
../build-mipsel-poky-linux/gdb/gdb: ELF 64-bit LSB shared object,
x86-64, version 1 (GNU/Linux), dynamically linked, interpreter
/opt/iro, BuildID[sha1]=7f985bbe4cb6c97558b159860b2498f6389b254e, for
GNU/Linux 3.2.0, not stripped
# ldd ../build-mipsel-poky-linux/gdb/gdb
../build-mipsel-poky-linux/gdb/gdb: /lib/x86_64-linux-gnu/libm.so.6:
version `GLIBC_2.29' not found (required by
../build-mipsel-poky-linux/gdb/gdb)
linux-vdso.so.1 => (0x00007fff8a0c2000)
...

# LD_LIBRARY_PATH=../recipe-sysroot-native/usr/libexec:../recipe-sysroot-native/usr/lib
../build-mipsel-poky-linux/gdb/gdb
bash: ../build-mipsel-poky-linux/gdb/gdb: No such file or directory

None of the techniques from my bag-of-tricks works.

I guess I could go grab the source myself, manually apply the patches
myself, build it, and see if that works.

Or I could sit down real hard and think about why I am trying to debug
the canadian-cross built tool on my development host... perhaps I
should try debugging the native (cross)-gdb on my native host. I'll
go try that now, but, in the mean time, I thought it was about time
for me to ask others for some clues.

Any clues or pointers?

--wpd


Re: [OE-core] oe-core recipe for defining directories in /

Mark Hatle
 

On 2/19/20 4:29 AM, Quentin Schulz wrote:
Hi JH,

On Wed, Feb 19, 2020 at 09:12:08PM +1100, JH wrote:
Hi,

Which recipe defines all directories in "/"? I need to make a bbapand
to add directories to /.
None. Or all of them, depends on how one sees it.
He is correct, this is shared between all package users. But with that said,
there is a base filesystem package and an associated 'syncing' file for all
other users.

The is a recipe called 'base-files', it builds up the core filesystem
infrastructure, as well as puts down a few basic files that all filesystems need.

The 'syncing' file is a corresponding file to used to coordinate permissions
between all recipes: meta/files/fs-perms.txt. fs-perms.txt is always consulted
when building any recipe to ensure that all common directory, permissions,
owners and groups get a common configuration.

If you are going to bbappend to base-files then you should also to add to
fs-perms.txt (you can do this, like a bbappend, by adding your own file to your
own layer and updating the variable 'FILESYSTEM_PERMS_TABLES' in one of your
global configuration files (usually your distro configuration file.)

but, this isn't the recommended way...

If the directories you are adding are custom to your product, I would recommend
creating a new package to manage this, as well as a custom
fs-perms.txt/FILESYSTEM_PERMS_TABLE entry to coordinate it between recipes that
use this new custom directory structure.

(If only one recipe uses that directory, then all of the overhead is simply not
needed! You only need this if multiple recipes are sharing a common directory
structure.)

--Mark

You just create a directory in do_install of a recipe.

You then make sure this directory is part of a package by checking it's
in one of the recipe's generated packages's FILES_<PACKAGE>.

Quentin


do_rootfs task took long time to finish

Marek Belisko
 

Hi,

I'm debugging strange issue that do_rootfs task took ~50 minutes. I
enabled buildstats and from that I learned it's quite long. Any ideas
how to debug this issue? My idea was to add timestamp to do_rootfs
tasks but I'm not sure if it's even possible. Thanks.

BR,

marek

--
as simple and primitive as possible
-------------------------------------------------
Marek Belisko - OPEN-NANDRA
Freelance Developer

Ruska Nova Ves 219 | Presov, 08005 Slovak Republic
Tel: +421 915 052 184
skype: marekwhite
twitter: #opennandra
web: http://open-nandra.com


Yocto Project Status WW07'20

Stephen Jolley
 

Current Dev Position: YP 3.1 M3

Next Deadline: YP 3.1 M3 build date 2/24/2020

 

Next Team Meetings:

 

Key Status/Updates:

  • The project recently updated its git hosting infrastructure and there were some issues encountered with the cgit http/https repository sharing. Those issues should now be resolved, apologies if they caused issues for anyone. The git:// protocol sharing was unaffected.
  • YP 3.0.2 rc2 is in QA with the report due soon.
  • We continue to see a small number of reproducibility issues with master which need resolving for green builds (in particular gstreamer and perl).
  • A significant memory usage issue was identified during bitbake parsing where memory usage would grow in each parser thread linearly per number of recipes parsed. This would therefore particularly affect large numbers of layers, multilibs and multiconfig. The fix has merged into bitbake along with the corresponding zeus and warrior branches. For one test case it reduced peak memory usage during parsing for 5 multiconfigs from 20GB to 2GB.
  • Warrior patches for 2.7.3 are out for review.
  • With the git infrastructure issue updated, we now have centos8 workers added to the autobuilder.
  • We are making various queued changes to the autobuilder configuration to fix bugs, improve efficiency and test coverage but this may result in some test result instability as we test and resolve issues.
  • We’re collecting a list of companies, products and projects which use the Yocto Project on the wiki: https://wiki.yoctoproject.org/wiki/Project_Users Please add any you know are missing (or email Richard/Stephen who can add).
  • The triage team is worried about attendance at triage meetings and the project is finding it hard to find people to help fix bugs. If anyone is willing to work on bugs, assistance would be greatly appreciated.

 

YP 3.1 Milestone Dates:

  • YP 3.1 M3 build date 2/24/2020
  • YP 3.1 M3 release date 3/6/2020
  • YP 3.1 M4 build date  3/30/2020
  • YP 3.1 M4 release date  4/24/2020

 

Planned upcoming dot releases:

  • YP 2.7.3 built and in QA
  • YP 2.7.3 release date 2/21/2020
  • YP 3.0.2 build date  2/3/2020
  • YP 3.0.2 release date 2/14/2020

 

Tracking Metrics:

 

The Yocto Project’s technical governance is through its Technical Steering Committee, more information is available at:

https://wiki.yoctoproject.org/wiki/TSC

 

The Status reports are now stored on the wiki at: https://wiki.yoctoproject.org/wiki/Weekly_Status

 

[If anyone has suggestions for other information you’d like to see on this weekly status update, let us know!]

 

Thanks,

 

Stephen K. Jolley

Yocto Project Program Manager

(    Cell:                (208) 244-4460

* Email:              sjolley.yp.pm@...

 


Opentk Support

Sheraz Ali
 

Hi,
Does anyone know how to enable opentk in yocto ( i.e is it available )

Thanks and Regards
Sheraz Ali Shah


Re: [OE-core] Yocto Project Status WW07'20

Alexander Kanavin
 

On Tue, 18 Feb 2020 at 17:10, <sjolley.yp.pm@...> wrote:
  • With the git infrastructure issue updated, we now have centos8 workers added to the autobuilder.
What are the current plans for centos 7? 

Alex


Re: [OE-core] [yocto] Change RO rootfs failed RF Kill Switch Status and Failed to start Run pending postinsts

JH
 

Hi Mikko,

On 2/18/20, Mikko.Rapeli@bmw.de <Mikko.Rapeli@bmw.de> wrote:
Well I have zeus and am using read-only rootfs with volatile binds and
I did not need anything extra. I would dig into this /var/log thing
and patch it away. I use systemd journal so no need for syslogs.
I actually not too worry about the rfkill error but more worried about
why the rfkill failed? Was it caused by some system problem? At the
moment, I could not get WiFi or LTE connected, which used to be
working in RW partition.

Thank you.

Kind regards,

- jh


oe-core recipe for defining directories in /

JH
 

Hi,

Which recipe defines all directories in "/"? I need to make a bbapand
to add directories to /.

Thank you.

Kind regards,

- jh


Re: [OE-core] Yocto Project Status WW07'20

Richard Purdie
 

On Tue, 2020-02-18 at 17:44 +0100, Alexander Kanavin wrote:
On Tue, 18 Feb 2020 at 17:10, <sjolley.yp.pm@gmail.com> wrote:
With the git infrastructure issue updated, we now have centos8
workers added to the autobuilder.
What are the current plans for centos 7?
Project members are keen to see support for centos7 retained.

I think we need to sort our buildtools plan, then we have more options
as we can use that on Centos7.

Cheers,

Richard


Re: [OE-core] [yocto] Change RO rootfs failed RF Kill Switch Status and Failed to start Run pending postinsts

JH
 

Hi Mikko,

On 2/18/20, Mikko.Rapeli@bmw.de <Mikko.Rapeli@bmw.de> wrote:
I think you may be missing volatile-binds package and service from your
image.
See poky/meta/recipes-core/volatile-binds/volatile-binds.bb
I got the source in my build system, it is zeus:

oe-core/meta/recipes-core/volatile-binds/volatile-binds.bb

./all-oe-linux/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/packages-split/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/sysroot-destdir/sysroot-providers/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata-pdata-input/runtime/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata-pdata-input/runtime-reverse/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata-pdata-input/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata/runtime/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata/runtime-reverse/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/pkgdata/volatile-binds
./all-oe-linux/volatile-binds/1.0-r0/license-destdir/volatile-binds

Are there correct?

It may be missing /var/log but with systemd there should not be needs to
write
to that location after image builds...
The volatile did not have the log, so /var/log -> volatile/log was an
invalid link, should I manually create it?

Thanks Mikko,

- jh

5381 - 5400 of 53882