Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

xxchan's official debug CI env #19696

Draft
wants to merge 20 commits into
base: main
Choose a base branch
from
5 changes: 0 additions & 5 deletions Makefile.toml
Original file line number Diff line number Diff line change
Expand Up @@ -287,11 +287,6 @@ do
exit 1
fi
done

if (( "$(du -sk ${PREFIX_LOG} | cut -f1)" > 4000 )) ; then
echo "$(tput setaf 1)ERROR: log size is significantly large ($(du -sh ${PREFIX_LOG} | cut -f1)).$(tput sgr0) Please disable unnecessary logs."
exit 1
fi
'''

[tasks.doc]
Expand Down
2 changes: 2 additions & 0 deletions ci/scripts/build-simulation.sh
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,8 @@ set -euo pipefail

source ci/scripts/common.sh

exit 88

echo "--- Generate RiseDev CI config"
cp ci/risedev-components.ci.env risedev-components.user.env

Expand Down
2 changes: 1 addition & 1 deletion ci/scripts/e2e-source-test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ python3 -m pip install --break-system-packages requests protobuf fastavro conflu
apt-get -y install jq

echo "--- e2e, inline test"
RUST_LOG="debug,risingwave_stream=info,risingwave_batch=info,risingwave_storage=info,risingwave_meta=info" \
RUST_LOG="debug,risingwave_stream=debug,risingwave_batch=info,risingwave_storage=info,risingwave_meta=info,events::stream::message::chunk=trace" \
risedev ci-start ci-inline-source-test
risedev slt './e2e_test/source_inline/**/*.slt' --keep-db-on-failure -j16
risedev slt './e2e_test/source_inline/**/*.slt.serial'
Expand Down
4 changes: 4 additions & 0 deletions ci/scripts/run-e2e-test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -81,11 +81,15 @@ chmod +x ./target/debug/risingwave_e2e_extended_mode_test
echo "--- e2e, $mode, streaming"
RUST_LOG="info,risingwave_stream=info,risingwave_batch=info,risingwave_storage=info,risingwave_stream::common::table::state_table=warn" \
cluster_start

exit 88

# Please make sure the regression is expected before increasing the timeout.
sqllogictest -p 4566 -d dev './e2e_test/streaming/**/*.slt' --junit "streaming-${profile}"
sqllogictest -p 4566 -d dev './e2e_test/backfill/sink/different_pk_and_dist_key.slt'

echo "--- Kill cluster"
exit 88
cluster_stop

echo "--- e2e, $mode, batch"
Expand Down
2 changes: 2 additions & 0 deletions ci/workflows/pull-request.yml
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,8 @@ auto-retry: &auto-retry
# Agent terminated because the AWS EC2 spot instance killed by AWS.
- signal_reason: agent_stop
limit: 3
manual:
permit_on_passed: true

cargo-cache: &cargo-cache
id: cargo
Expand Down
6 changes: 3 additions & 3 deletions e2e_test/source_inline/kafka/issue_19563.slt.serial
Original file line number Diff line number Diff line change
Expand Up @@ -46,9 +46,9 @@ select array_length(upstream_fragment_ids) from rw_fragments where array_contain
----
3

# XXX: wait until source reader is ready. then produce data.
# This is a temporary workaround for a data loss bug https://github.com/risingwavelabs/risingwave/issues/19681#issuecomment-2532183002
sleep 2s
# # XXX: wait until source reader is ready. then produce data.
# # This is a temporary workaround for a data loss bug https://github.com/risingwavelabs/risingwave/issues/19681#issuecomment-2532183002
# sleep 2s

system ok
cat <<EOF | rpk topic produce test-topic-19563
Expand Down
12 changes: 6 additions & 6 deletions e2e_test/source_inline/kafka/nexmark.slt
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
include ../../nexmark/produce_kafka.slt.part
include ../../nexmark/create_sources_kafka.slt.part
# include ../../nexmark/produce_kafka.slt.part
# include ../../nexmark/create_sources_kafka.slt.part

control substitution off
# control substitution off

include ../../streaming/nexmark/create_views.slt.part
include ../../streaming/nexmark/test_mv_result.slt.part
# include ../../streaming/nexmark/create_views.slt.part
# include ../../streaming/nexmark/test_mv_result.slt.part

include ../../nexmark/drop_sources_kafka.slt.part
# include ../../nexmark/drop_sources_kafka.slt.part
54 changes: 27 additions & 27 deletions e2e_test/source_inline/kafka/tpch.slt
Original file line number Diff line number Diff line change
@@ -1,30 +1,30 @@
include ../../tpch/produce_kafka.slt.part
include ../../tpch/create_sources_kafka.slt.part
# include ../../tpch/produce_kafka.slt.part
# include ../../tpch/create_sources_kafka.slt.part

control substitution off
# control substitution off

include ../../streaming/tpch/create_views.slt.part
include ../../streaming/tpch/q1.slt.part
include ../../streaming/tpch/q2.slt.part
include ../../streaming/tpch/q3.slt.part
include ../../streaming/tpch/q4.slt.part
include ../../streaming/tpch/q5.slt.part
include ../../streaming/tpch/q6.slt.part
include ../../streaming/tpch/q7.slt.part
include ../../streaming/tpch/q8.slt.part
include ../../streaming/tpch/q9.slt.part
include ../../streaming/tpch/q10.slt.part
include ../../streaming/tpch/q11.slt.part
include ../../streaming/tpch/q12.slt.part
include ../../streaming/tpch/q13.slt.part
include ../../streaming/tpch/q14.slt.part
include ../../streaming/tpch/q15.slt.part
include ../../streaming/tpch/q16.slt.part
include ../../streaming/tpch/q17.slt.part
include ../../streaming/tpch/q18.slt.part
include ../../streaming/tpch/q19.slt.part
include ../../streaming/tpch/q20.slt.part
include ../../streaming/tpch/q21.slt.part
include ../../streaming/tpch/q22.slt.part
# include ../../streaming/tpch/create_views.slt.part
# include ../../streaming/tpch/q1.slt.part
# include ../../streaming/tpch/q2.slt.part
# include ../../streaming/tpch/q3.slt.part
# include ../../streaming/tpch/q4.slt.part
# include ../../streaming/tpch/q5.slt.part
# include ../../streaming/tpch/q6.slt.part
# include ../../streaming/tpch/q7.slt.part
# include ../../streaming/tpch/q8.slt.part
# include ../../streaming/tpch/q9.slt.part
# include ../../streaming/tpch/q10.slt.part
# include ../../streaming/tpch/q11.slt.part
# include ../../streaming/tpch/q12.slt.part
# include ../../streaming/tpch/q13.slt.part
# include ../../streaming/tpch/q14.slt.part
# include ../../streaming/tpch/q15.slt.part
# include ../../streaming/tpch/q16.slt.part
# include ../../streaming/tpch/q17.slt.part
# include ../../streaming/tpch/q18.slt.part
# include ../../streaming/tpch/q19.slt.part
# include ../../streaming/tpch/q20.slt.part
# include ../../streaming/tpch/q21.slt.part
# include ../../streaming/tpch/q22.slt.part

include ../../tpch/drop_sources_kafka.slt.part
# include ../../tpch/drop_sources_kafka.slt.part
3 changes: 2 additions & 1 deletion src/meta/src/manager/diagnose.rs
Original file line number Diff line number Diff line change
Expand Up @@ -171,7 +171,8 @@ impl DiagnoseCommand {
&mut row,
worker_node.get_state().ok().map(|s| s.as_str_name()),
);
row.add_cell(worker_node.parallelism().into());
// FIXME:
// row.add_cell(worker_node.parallelism().into());
try_add_cell(
&mut row,
worker_node.property.as_ref().map(|p| p.is_streaming),
Expand Down
10 changes: 5 additions & 5 deletions src/utils/runtime/src/logger.rs
Original file line number Diff line number Diff line change
Expand Up @@ -289,15 +289,15 @@ pub fn init_risingwave_logger(settings: LoggerSettings) {
});

let fmt_layer = match deployment {
Deployment::Ci => fmt_layer
.compact()
.with_filter(FilterFn::new(|metadata| metadata.is_event())) // filter-out all span-related info
.boxed(),
// Deployment::Ci => fmt_layer
// .compact()
// .with_filter(FilterFn::new(|metadata| metadata.is_event())) // filter-out all span-related info
// .boxed(),
Deployment::Cloud => fmt_layer
.json()
.map_event_format(|e| e.with_current_span(false)) // avoid duplication as there's a span list field
.boxed(),
Deployment::Other => {
Deployment::Ci | Deployment::Other => {
if env_var_is_true("ENABLE_PRETTY_LOG") {
fmt_layer.pretty().boxed()
} else {
Expand Down
Loading