由于未找到 ext4,数据流 (Beam 2.12) 未启动

Dataflow (Beam 2.12) does not start due to ext4 not found

我在 运行 数据流作业 (Beam 2.12) 时看到所有类型的奇怪错误。

作业基本上从 pubsub 获取输入,read/writes from/to Datastore 将结果写入 pubsub。

几个警告 W 和错误 E 出现在 Stackdriver 日志中。目前尚不清楚如何解决这些问题。到目前为止,我们使用的是 Beam 2.9,没有遇到任何这些问题。

下面提供了部分(编辑过的)日志转储。

W  acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. 
W  ACPI: PCI Interrupt Link [LNKC] enabled at IRQ 11 
W  ACPI: PCI Interrupt Link [LNKD] enabled at IRQ 10 
W  i8042: Warning: Keylock active 
W  GPT:Primary header thinks Alt. header is not at the end of the disk. 
W  GPT:52428799 != 62914559 
W  GPT:Alternate GPT header not at the end of the disk. 
W  GPT:52428799 != 62914559 
W  GPT: Use GNU Parted to correct GPT errors. 
W  device-mapper: verity: Argument 0: 'payload=PARTUUID=XXX' 
W  device-mapper: verity: Argument 1: 'hashtree=PARTUUID=XXX' 
W  device-mapper: verity: Argument 2: 'hashstart=2539520' 
W  device-mapper: verity: Argument 3: 'alg=sha1' 
W  device-mapper: verity: Argument 4: 'root_hexdigest=900...' 
W  device-mapper: verity: Argument 5: 'salt=b113702...' 
W  [/usr/lib/tmpfiles.d/var.conf:12] Duplicate line for path "/var/run", ignoring. 
W  Cannot set file attribute for '/var/log/journal', value=0x00800000, mask=0x00800000: Operation not supported 
W  Cannot set file attribute for '/var/log/journal/2a0c3f0af65e8318a0b8f3eb...', value=0x00800000, mask=0x00800000: Operation not supported 
W  Could not load the device policy file. 
W  [WARNING:persistent_integer.cc(96)] cannot open /var/lib/metrics/version.cycle for reading: No such file or directory 
W  WARNING Could not update the authorized keys file for user root. [Errno 30] Read-only file system: '/root/.ssh'. 
W  [CLOUDINIT] cc_write_files.py[WARNING]: Undecodable permissions None, assuming 420 
...
E  Error initializing dynamic plugin prober: Error (re-)creating driver directory: mkdir /usr/libexec/kubernetes: read-only file system 
W  No api server defined - no node status update will be sent. 
W  Failed to retrieve checkpoint for "kubelet_internal_checkpoint": checkpoint is not found 
W  Unknown healthcheck type 'NONE' (expected 'CMD') in container 7df5acdbd1ad6756e3e409c6e8760d274bdc03f83bf... 
E  while reading 'google-dockercfg-url' metadata: http status code: 404 while fetching url http://metadata.google.internal./computeMetadata/v1/instance/attributes/google-dockercfg-url 
E  while reading 'google-dockercfg' metadata: http status code: 404 while fetching url http://metadata.google.internal./computeMetadata/v1/instance/attributes/google-dockercfg 
W  Unknown healthcheck type 'NONE' (expected 'CMD') in container 3aa8c92a0b7d746d7004768d5182f0558a0c0c90dfcd5... 
W  Unknown healthcheck type 'NONE' (expected 'CMD') in container 64b9fb0459f88833dee78943c32598761154e4a49d708... 
W  Unknown healthcheck type 'NONE' (expected 'CMD') in container d2edf1c5e89b746e8c9c96b2a39a9d7ac7da2ecf52f96d...
W  Unknown healthcheck type 'NONE' (expected 'CMD') in container b2448a8792ad63059bb70f1f6f12385caae7a833018d05...
E  EXT4-fs (sdb): VFS: Can't find ext4 filesystem 
E  Error syncing pod c386113... ("dataflow-...-harness-z656_default(c386113...)"), skipping: failed to "StartContainer" for "java-streaming" with CrashLoopBackOff: "Back-off 10s restarting failed container=java-streaming pod=dataflow-...-harness-z656_default(c386113...)" 
W  [WARNING:metrics_daemon.cc(619)] cannot read /sys/devices/system/cpu/cpu0/cpufreq/cpuinfo_max_freq 
E  EXT4-fs (sdd): VFS: Can't find ext4 filesystem 
...
W  Unknown healthcheck type 'NONE' (expected 'CMD') in container 675eb66a9e794b3dea03b62c3bdaf539034c998bf11c... 
E  Error syncing pod c386113... ("dataflow-...-harness-z656_default(c386113...)"), skipping: failed to "StartContainer" for "java-streaming" with CrashLoopBackOff: "Back-off 40s restarting failed container=java-streaming pod=dataflow-...-harness-z656_default(c386113...)" 
E  Error syncing pod c386113... ("dataflow-...-harness-z656_default(c386113...)"), skipping: failed to "StartContainer" for "java-streaming" with CrashLoopBackOff: "Back-off 40s restarting failed container=java-streaming pod=dataflow-...-harness-z656_default(c386113...)" 
E  Error syncing pod c386113... ("dataflow-...-harness-z656_default(c386113...)"), skipping: failed to "StartContainer" for "java-streaming" with CrashLoopBackOff: "Back-off 40s restarting failed container=java-streaming pod=dataflow-...-harness-z656_default(c386113...)" 
W  Unknown healthcheck type 'NONE' (expected 'CMD') in container 7d7536b93cb92562bdd12da3fd25a53caea8c9a9e1cee603b3999dfdd5681a27 
E  Error syncing pod c386113... ("dataflow-...-harness-z656_default(c386113...)"), skipping: failed to "StartContainer" for "java-streaming" with CrashLoopBackOff: "Back-off 1m20s restarting failed container=java-streaming pod=dataflow-...-harness-z656_default(c386113...)" 

我通过升级几个依赖项解决了这个问题。

maven 版本插件帮助我做到了这一点,我通过将以下内容添加到我的 .pom 文件来安装插件:

<plugin>
  <groupId>org.codehaus.mojo</groupId>
  <artifactId>versions-maven-plugin</artifactId>
  <version>2.5</version>
</plugin>

然后我查了库更新,更新了。如前所述,我怀疑它是一个较旧的 BigTable 客户端 here

mvn versions:display-dependency-updates
mvn versions:use-latest-versions