d4c1b39d03
[fix](multi table) restrict the multi tables load memory under high concurrency with a large number of tables ( #39992 ) ( #41131 )
...
pick (#39992 )
BE node was killed by OOM-killer when use multi table load under high
concurrency with a large number of tables(128 concurrency and every
concurrency load 200 tables).
This pr restricts the multi tables load memory under this issue. If
memory reaches hard limit, new task will be rejected and return
directly.
2024-09-24 16:34:32 +08:00
4237341e0d
[fix](move-memtable) do not execute close if create rowset failed when loading MOW table ( #40105 ) ( #41132 )
...
pick (#40105 )
Core dump happened when load to MOW table:
```
Check failure stack trace: ***
@ 0x55fae437d246 google::LogMessage::SendToLog()
@ 0x55fae4379c90 google::LogMessage::Flush()
@ 0x55fae437da89 google::LogMessageFatal::~LogMessageFatal()
@ 0x55faacf26bbf doris::BaseTablet::check_delete_bitmap_correctness()
@ 0x55fab05049ef doris::RowsetBuilder::commit_txn()
@ 0x55fab09026e8 doris::LoadStreamWriter::close()
@ 0x55fab089eff7 std::_Function_handler<>::_M_invoke()
@ 0x55fab0d14d7c doris::WorkThreadPool<>::work_thread()
@ 0x55fae76ae6f0 execute_native_thread_routine
@ 0x7fa32ea45ac3 (unknown)
@ 0x7fa32ead7850 (unknown)
@ (nil) (unknown)
Query id: a21981d5c8ef4113-84df9a5a8680e004 ***
is nereids: 0 ***
tablet id: 0 ***
Aborted at 1724668499 (unix time) try "date -d @1724668499" if you are using GNU date ***
Current BE git commitID: 2f848737c1 ***
SIGABRT unknown detail explain (@0x20db) received by PID 8411 (TID 9837 OR 0x7f9e42cfe640) from PID 8411; stack trace: ***
0# doris::signal::(anonymous namespace)::FailureSignalHandler(int, siginfo_t*, void*) at /home/zcp/repo_center/doris_master/doris/be/src/common/signal_handler.h:421
1# 0x00007FA32E9F3520 in /lib/x86_64-linux-gnu/libc.so.6
2# pthread_kill at ./nptl/pthread_kill.c:89
3# raise at ../sysdeps/posix/raise.c:27
4# abort at ./stdlib/abort.c:81
5# 0x000055FAE4387B1D in /mnt/hdd01/ci/master-deploy/be/lib/doris_be
6# 0x000055FAE437A15A in /mnt/hdd01/ci/master-deploy/be/lib/doris_be
7# google::LogMessage::SendToLog() in /mnt/hdd01/ci/master-deploy/be/lib/doris_be
8# google::LogMessage::Flush() in /mnt/hdd01/ci/master-deploy/be/lib/doris_be
9# google::LogMessageFatal::~LogMessageFatal() in /mnt/hdd01/ci/master-deploy/be/lib/doris_be
10# doris::BaseTablet::check_delete_bitmap_correctness(std::shared_ptr, long, long, std::unordered_set, std::equal_to, std::allocator > const&, std::vector, std::allocator > >*) at /home/zcp/repo_center/doris_master/doris/be/src/olap/base_tablet.cpp:1152
11# doris::RowsetBuilder::commit_txn() at /home/zcp/repo_center/doris_master/doris/be/src/olap/rowset_builder.cpp:316
12# doris::LoadStreamWriter::close() at /home/zcp/repo_center/doris_master/doris/be/src/runtime/load_stream_writer.cpp:311
13# std::_Function_handler::_M_invoke(std::_Any_data const&) at /var/local/ldb-toolchain/bin/../lib/gcc/x86_64-linux-gnu/11/../../../../include/c++/11/bits/std_function.h:291
14# doris::WorkThreadPool::work_thread(int) at /home/zcp/repo_center/doris_master/doris/be/src/util/work_thread_pool.hpp:159
15# execute_native_thread_routine at ../../../../../libstdc+-v3/src/c+11/thread.cc:84
16# start_thread at ./nptl/pthread_create.c:442
17# 0x00007FA32EAD7850 at ../sysdeps/unix/sysv/linux/x86_64/clone3.S:83
```
If create rowset failed,` calc_delete_bitmap_task` still could be
executed:
```
add segment failed load_id=5649413b98976f0d-a105b42749f561b0, txn_id=2, tablet_id=10088, status=[INTERNAL_ERROR]create row
set failed
...
submit calc delete bitmap task to executor, tablet_id: 10088, txn_id: 2
```
This PR skips close to avoid `submit_calc_delete_bitmap_task` if create
rowset failed when loading MOW table to solve this problem.
2024-09-24 16:02:22 +08:00
602275d1fe
[improvement](stream load) display the time of receiving network data ( #40735 ) ( #41139 )
...
pick (#40735 )
Display the time of receiving network data.
For example:
```
Stream load result: {
"TxnId": 19,
"Label": "c10ee032-c19c-484f-80b2-16f6c1d0c209",
"Comment": "",
"TwoPhaseCommit": "false",
"Status": "Success",
"Message": "OK",
"NumberTotalRows": 6,
"NumberLoadedRows": 6,
"NumberFilteredRows": 0,
"NumberUnselectedRows": 0,
"LoadBytes": 60,
"LoadTimeMs": 32,
"BeginTxnTimeMs": 0,
"StreamLoadPutTimeMs": 4,
"ReadDataTimeMs": 0,
"WriteDataTimeMs": 4,
"ReceiveDataTimeMs": 6,
"CommitAndPublishTimeMs": 22
}
```
2024-09-24 16:00:28 +08:00
96b2fa1a24
[fix](move-memtable) fix bvar g_load_stream_file_writer_cnt ( #39075 ) ( #40985 )
...
backport #39075
2024-09-20 20:20:49 +08:00
8e860a26a7
[fix](systable) fix unstable case for partitions table ( #40553 ) ( #41043 )
...
bp #40553
2024-09-20 17:13:30 +08:00
549bc3e288
[fix](pipelinex) fix fragment instance progress reports ( #40325 ) ( #40987 )
...
backport #40325
2024-09-19 23:58:38 +08:00
a6e2364f56
[cherrypick](branch21) pick some exception handleing pr in page io ( #40883 )
...
## Proposed changes
pick
https://github.com/apache/doris/pull/39536
https://github.com/apache/doris/pull/39913
https://github.com/apache/doris/pull/39964
https://github.com/apache/doris/pull/40020
---------
Co-authored-by: yiguolei <yiguolei@gmail.com >
2024-09-16 08:56:43 +08:00
b52b572ade
[branch-2.1](memory) When Load ends, check memory tracker value returns is equal to 0 ( #40850 )
...
pick
#38960
#39908
#40043
#40092
#40016
#40439
---------
Co-authored-by: hui lai <1353307710@qq.com >
Co-authored-by: yiguolei <676222867@qq.com >
2024-09-15 23:47:53 +08:00
f16615a1fc
[branch-2.1](memory) Allocator support address sanitizers ( #40836 )
...
pick
#33396
#33862
#33853
#33732
#33841
#33933
#34901
#35014
---------
Co-authored-by: yiguolei <676222867@qq.com >
Co-authored-by: yiguolei <yiguolei@gmail.com >
2024-09-14 12:12:44 +08:00
7851563829
[fix](brpc_client_cache) resolve hostname in DNS cache before passing to brpc ( #40074 ) ( #40786 )
...
backport #40074
2024-09-13 14:28:01 +08:00
3395cd5ce9
[PipelineX](improvement) Prepare tasks in parallel ( #40270 )
...
## Proposed changes
Issue Number: close #xxx
<!--Describe your changes.-->
2024-09-13 13:34:29 +08:00
3604d63184
[Branch 2.1] backport systable PR (#34384,#40153,#40456,#40455,#40568) ( #40687 )
...
backport
https://github.com/apache/doris/pull/40568
https://github.com/apache/doris/pull/40455
https://github.com/apache/doris/pull/40456
https://github.com/apache/doris/pull/40153
https://github.com/apache/doris/pull/34384
Test result:
2024-09-11 11:00:45.618 INFO [suite-thread-1] (SuiteContext.groovy:309)
- Recover original connection
2024-09-11 11:00:45.619 INFO [suite-thread-1] (Suite.groovy:359) -
Execute sql: REVOKE SELECT_PRIV ON
test_partitions_schema_db.duplicate_table FROM partitions_user
2024-09-11 11:00:45.625 INFO [suite-thread-1] (SuiteContext.groovy:299)
- Create new connection for user 'partitions_user'
2024-09-11 11:00:45.632 INFO [suite-thread-1] (Suite.groovy:1162) -
Execute tag: select_check_5, sql: select
TABLE_CATALOG,TABLE_SCHEMA,TABLE_NAME,PARTITION_NAME,SUBPARTITION_NAME,PARTITION_ORDINAL_POSITION,SUBPARTITION_ORDINAL_POSITION,PARTITION_METHOD,SUBPARTITION_METHOD,PARTITION_EXPRESSION,SUBPARTITION_EXPRESSION,PARTITION_DESCRIPTION,TABLE_ROWS,AVG_ROW_LENGTH,DATA_LENGTH,MAX_DATA_LENGTH,INDEX_LENGTH,DATA_FREE,CHECKSUM,PARTITION_COMMENT,NODEGROUP,TABLESPACE_NAME
from information_schema.partitions where
table_schema="test_partitions_schema_db" order by
TABLE_CATALOG,TABLE_SCHEMA,TABLE_NAME,PARTITION_NAME,SUBPARTITION_NAME,PARTITION_ORDINAL_POSITION,SUBPARTITION_ORDINAL_POSITION,PARTITION_METHOD,SUBPARTITION_METHOD,PARTITION_EXPRESSION,SUBPARTITION_EXPRESSION,PARTITION_DESCRIPTION,TABLE_ROWS,AVG_ROW_LENGTH,DATA_LENGTH,MAX_DATA_LENGTH,INDEX_LENGTH,DATA_FREE,CHECKSUM,PARTITION_COMMENT,NODEGROUP,TABLESPACE_NAME
2024-09-11 11:00:45.644 INFO [suite-thread-1] (SuiteContext.groovy:309)
- Recover original connection
2024-09-11 11:00:45.645 INFO [suite-thread-1] (ScriptContext.groovy:120)
- Run test_partitions_schema in
/root/doris/workspace/doris/regression-test/suites/query_p0/system/test_partitions_schema.groovy
succeed
2024-09-11 11:00:45.652 INFO [main] (RegressionTest.groovy:259) - Start
to run single scripts
2024-09-11 11:01:10.321 INFO [main] (RegressionTest.groovy:380) -
Success suites:
/root/doris/workspace/doris/regression-test/suites/query_p0/system/test_partitions_schema.groovy:
group=default,p0, name=test_partitions_schema
2024-09-11 11:01:10.322 INFO [main] (RegressionTest.groovy:459) - All
suites success.
____ _ ____ ____ _____ ____
| _ \ / \ / ___/ ___|| ____| _ \
| |_) / _ \ \___ \___ \| _| | | | |
| __/ ___ \ ___) |__) | |___| |_| |
|_| /_/ \_\____/____/|_____|____/
2024-09-11 11:01:10.322 INFO [main] (RegressionTest.groovy:410) - Test 1
suites, failed 0 suites, fatal 0 scripts, skipped 0 scripts
2024-09-11 11:01:10.322 INFO [main] (RegressionTest.groovy:119) - Test
finished
2024-09-11 11:03:00.712 INFO [suite-thread-1] (Suite.groovy:1162) -
Execute tag: select_check_5, sql: select * from
information_schema.table_options ORDER BY
TABLE_CATALOG,TABLE_SCHEMA,TABLE_NAME,TABLE_MODEL,TABLE_MODEL_KEY,DISTRIBUTE_KEY,DISTRIBUTE_TYPE,BUCKETS_NUM,PARTITION_NUM;
2024-09-11 11:03:00.729 INFO [suite-thread-1] (SuiteContext.groovy:309)
- Recover original connection
2024-09-11 11:03:00.731 INFO [suite-thread-1] (ScriptContext.groovy:120)
- Run test_table_options in
/root/doris/workspace/doris/regression-test/suites/query_p0/system/test_table_options.groovy
succeed
2024-09-11 11:03:04.817 INFO [main] (RegressionTest.groovy:259) - Start
to run single scripts
2024-09-11 11:03:28.741 INFO [main] (RegressionTest.groovy:380) -
Success suites:
/root/doris/workspace/doris/regression-test/suites/query_p0/system/test_table_options.groovy:
group=default,p0, name=test_table_options
2024-09-11 11:03:28.742 INFO [main] (RegressionTest.groovy:459) - All
suites success.
____ _ ____ ____ _____ ____
| _ \ / \ / ___/ ___|| ____| _ \
| |_) / _ \ \___ \___ \| _| | | | |
| __/ ___ \ ___) |__) | |___| |_| |
|_| /_/ \_\____/____/|_____|____/
2024-09-11 11:03:28.742 INFO [main] (RegressionTest.groovy:410) - Test 1
suites, failed 0 suites, fatal 0 scripts, skipped 0 scripts
2024-09-11 11:03:28.742 INFO [main] (RegressionTest.groovy:119) - Test
finished
*************************** 7. row ***************************
PartitionId: 18035
PartitionName: p100
VisibleVersion: 2
VisibleVersionTime: 2024-09-11 10:59:28
State: NORMAL
PartitionKey: col_1
Range: [types: [INT]; keys: [83647]; ..types: [INT]; keys: [2147483647];
)
DistributionKey: pk
Buckets: 10
ReplicationNum: 1
StorageMedium: HDD
CooldownTime: 9999-12-31 15:59:59
RemoteStoragePolicy:
LastConsistencyCheckTime: NULL
DataSize: 2.872 KB
IsInMemory: false
ReplicaAllocation: tag.location.default: 1
IsMutable: true
SyncWithBaseTables: true
UnsyncTables: NULL
CommittedVersion: 2
RowCount: 4
7 rows in set (0.01 sec)
---------
Co-authored-by: Mingyu Chen <morningman.cmy@gmail.com >
2024-09-12 11:50:09 +08:00
354967c09f
[branch-2.1](memory) pick reserve memory and workload group ( #40543 )
...
1. pick #38494
2. pick #39862
3. remove vdata_stream_test, master has been removed
2024-09-09 21:16:06 +08:00
facce8b4d5
[fix](move-memtable) multi replica tables should tolerate minority failures ( #38003 ) ( #40477 )
...
backport #38003
2024-09-09 11:30:46 +08:00
d9fa59be4d
[Chore](runtime-filter) avoid dcheck fail when rf merge failed ( #39172 ) ( #40409 )
...
pick from #39172
2024-09-05 14:50:47 +08:00
67a2daf33c
[fix](execution) fix wrong check when blocking result sink ( #40289 )
...
Otherwise, when client stop fetch result data, the result sink will
still adding batch to the result queue,
which causing BE OOM.
2024-09-03 14:17:42 +08:00
d949df0c4f
[chore](cancel) Make enable_pipeline_task_leakage_detect false by default #40145 ( #40192 )
...
cherry pick from #40145
2024-08-30 22:48:00 +08:00
090474a815
[branch-2.1](load) Lower lock granularity to reduce loading CPU pressure ( #40135 )
...
pick https://github.com/apache/doris/pull/40134
2024-08-29 23:22:58 +08:00
c70d7d8423
[fix](sink) Do not block result sink on pipeline engine ( #40094 )
...
## Proposed changes
Issue Number: close #xxx
<!--Describe your changes.-->
2024-08-29 14:04:25 +08:00
af6c39d770
[fix](compile) compile fail on MacOS because 'bits/types/struct_timespec.h' not found ( #39567 ) ( #40044 )
...
pick #39567 to branch-2.1
Co-authored-by: 神技圈子 <songguangfan@gmail.com >
Co-authored-by: 宋光璠 <songguangfan@sf.com >
2024-08-28 22:01:49 +08:00
357394bb3e
[branch-2.1]Reset io limit default value ( #39898 )
...
pick #39842
2024-08-26 14:27:35 +08:00
2dea859bdb
[debug](rpc) debug rpc time consumption problem ( #39852 )
...
## Proposed changes
Issue Number: close #xxx
Add detail RPC time info for each channel, sorted by max rpc time of
channels:
```
DATA_STREAM_SINK_OPERATOR (id=1,dst_id=1):
- Partitioner: Crc32HashPartitioner(64)
- BlocksProduced: 74
- BrpcSendTime: 2.689us
- BrpcSendTime.Wait: 0ns
- BytesSent: 89.35 KB
- CloseTime: 680.152us
- CompressTime: 0ns
- ExecTime: 160.663ms
- InitTime: 263.608us
- InputRows: 32.512K (32512)
- LocalBytesSent: 0.00
- LocalSendTime: 0ns
- LocalSentRows: 0
- MemoryUsage:
- PeakMemoryUsage: 80.00 KB
- MergeBlockTime: 0ns
- OpenTime: 4.113ms
- OverallThroughput: 0.0 /sec
- PendingFinishDependency: 41.179ms
- RowsProduced: 32.512K (32512)
- RpcAvgTime: 11.850ms
- RpcCount: 10
- RpcMaxTime: 86.891ms
- RpcMinTime: 15.200ms
- RpcSumTime: 118.503ms
- SerializeBatchTime: 13.517ms
- SplitBlockDistributeByChannelTime: 38.923ms
- SplitBlockHashComputeTime: 2.659ms
- UncompressedRowBatchSize: 135.19 KB
- WaitForDependencyTime: 0ns
- WaitForRpcBufferQueue: 0ns
RpcInstanceDetails:
- Instance 85d4f75b72a9ea61: Count: 4, MaxTime: 36.238ms, MinTime: 12.107ms, AvgTime: 21.722ms, SumTime: 86.891ms
- Instance 85d4f75b72a9ea91: Count: 3, MaxTime: 11.107ms, MinTime: 2.431ms, AvgTime: 5.470ms, SumTime: 16.412ms
- Instance 85d4f75b72a9eac1: Count: 3, MaxTime: 7.554ms, MinTime: 3.160ms, AvgTime: 5.066ms, SumTime: 15.200m
```
2024-08-24 19:59:39 +08:00
ae4d747c13
[branch-2.1](memory) Modify memory gc conf and add crash_in_alloc_large_memory_bytes ( #39834 )
...
pick #39611
2024-08-24 09:21:35 +08:00
37443aa7e1
[improve](move-memtable) reuse connection in load_stream_stub ( #39231 ) ( #39762 )
...
backport #39231
2024-08-23 22:46:28 +08:00
6ceb574aa0
[branch-2.1]Pick IO limit/workload group usage table ( #39839 )
2024-08-23 18:51:47 +08:00
1367f74e7a
[branch-2.1](memory) Optimize ClearCacheActionimplementation ( #39796 )
...
pick #38438
2024-08-23 01:51:14 +08:00
06a0b35704
[chore] Comment for tv_nsec ( #39752 )
...
just comment.
2024-08-22 22:16:21 +08:00
13b882a4cc
[branch-2.1](memory) Add memory metrics to bvar ( #39763 )
...
pick #38391
2024-08-22 17:34:30 +08:00
8ce8887b75
[branch-2.1](memory) Refactor refresh workload groups weighted memory ratio and record refresh interval memory growth ( #39760 )
...
pick #38168
overwrites changes in #37221 on workload_group_manager.cpp. If need to
pick 37221, ignore it.
2024-08-22 17:33:11 +08:00
56cc9cc304
[fix](cancel)) Fix pipeline task leak cancel ( #39697 )
...
pick #39737
2024-08-22 00:40:22 +08:00
610f69432a
[improvement](segmentcache) limit segment cache by fd limit or memory… ( #39689 )
...
… (#39658 )
remove a useless config.
2024-08-21 15:19:52 +08:00
0bfcee1251
[opt](file-cache) support system table file_cache_statistics ( #39552 )
...
1. Add new system table: `file_cache_statistics`
This table is used for viewing metrics related to file cache on BE side
```
mysql> select * from information_schema.file_cache_statistics limit 10;
+-------+---------------+----------------------------+--------------------------------+--------------------+
| BE_ID | BE_IP | CACHE_PATH | METRIC_NAME | METRIC_VALUE |
+-------+---------------+----------------------------+--------------------------------+--------------------+
| 10003 | 172.20.32.136 | /mnt/output/be/file_cache/ |
disposable_queue_curr_elements | 0 |
| 10003 | 172.20.32.136 | /mnt/output/be/file_cache/ |
disposable_queue_curr_size | 0 |
| 10003 | 172.20.32.136 | /mnt/output/be/file_cache/ |
disposable_queue_max_elements | 102400 |
| 10003 | 172.20.32.136 | /mnt/output/be/file_cache/ |
disposable_queue_max_size | 21474836480 |
| 10003 | 172.20.32.136 | /mnt/output/be/file_cache/ | hits_ratio |
0.8539634687001242 |
| 10003 | 172.20.32.136 | /mnt/output/be/file_cache/ | hits_ratio_1h | 0
|
| 10003 | 172.20.32.136 | /mnt/output/be/file_cache/ | hits_ratio_5m | 0
|
| 10003 | 172.20.32.136 | /mnt/output/be/file_cache/ |
index_queue_curr_elements | 0 |
| 10003 | 172.20.32.136 | /mnt/output/be/file_cache/ |
index_queue_curr_size | 0 |
| 10003 | 172.20.32.136 | /mnt/output/be/file_cache/ |
index_queue_max_elements | 102400 |
+-------+---------------+----------------------------+--------------------------------+--------------------+
```
It will show metrics of file caches on each BE.
2. Add new metrics `hits_ratio_1h` and `hits_ratio_5m` for file cache
This 2 metrics will show the hit ratio of file cache in recent 1 hour or
5 minutes.
So that we can know recent hit ratio instead of global historical hit
ratio.
2024-08-21 10:03:39 +08:00
830f250a80
[opt](query cancel) cancel query if it has pipeline task leakage #39223 ( #39537 )
...
pick #39223 with some modifications. Optimization will only be applied
to pipeline x.
2024-08-19 14:33:59 +08:00
eea3676791
[fix](group commit) fix group commit insert rpc may stuck ( #39391 ) ( #39458 )
...
pick https://github.com/apache/doris/pull/39391
2024-08-16 13:19:00 +08:00
6f1d9812bb
[Bug](brpc) fix sync_filter_size/apply_filterv2 has wrong closure ( #39299 )
...
pick from #39155
2024-08-13 19:01:22 +08:00
1ef42dd94b
[Fix](load) The value of the index id printed in the log is incorrect #38790 ( #39131 )
...
cherry pick from #38790
2024-08-09 12:31:33 +08:00
0571342538
[fix](sink) The issue with 2GB limit of protocol buffer ( #37990 ) ( #39112 )
...
```
Fail to serialize doris.PFetchDataResult
```
If the size of `PFetchDataResult` is greater than 2G, protocol buffer
cannot serialize the message.
pick #37990
2024-08-09 04:01:56 +08:00
2ec1a6a7e7
[fix](group commit) Modify group commit commit/abort txn timeout as stream load ( #39003 ) ( #39069 )
...
pick https://github.com/apache/doris/pull/39003
2024-08-08 14:36:29 +08:00
749c9f7b56
[fix](group commit) fix repaly wal check label status ( #38883 ) ( #38997 )
...
pick https://github.com/apache/doris/pull/38883
2024-08-07 22:06:59 +08:00
3abb222064
[fix](group commit) Fix test_group_commit_async_wal_msg_fault_injection case ( #35313 ) ( #38911 )
...
pick https://github.com/apache/doris/pull/35313
2024-08-06 17:57:22 +08:00
8fa0710cb3
[branch-2.1](load) fix miss writer in concurrency incremental open ( #38605 ) ( #38793 )
...
pick https://github.com/apache/doris/pull/38605
2024-08-05 08:56:23 +08:00
e8690b62ee
[fix](group commit) Pick add debug log show why group commit not work; delete wal when replay success ( #38611 ) ( #38659 )
...
Pick https://github.com/apache/doris/pull/38611
2024-08-01 16:59:54 +08:00
182bf4d323
[chore](fe) Returns dropped tables in GetMeta request ( #38541 )
...
Cherry-pick #38019
2024-07-31 10:57:00 +08:00
017dad8c54
[fix](type)support runtime predicate for time type ( #38258 ) ( #38465 )
...
## Proposed changes
https://github.com/apache/doris/pull/38258
Issue Number: close #xxx
<!--Describe your changes.-->
2024-07-31 10:27:36 +08:00
e9f12fac47
[fix](load) fix no error url for stream load #38325 ( #38417 )
...
cherry pick from #38325
2024-07-28 19:06:57 +08:00
91f0301b43
[fix](group commit) Pick some group commit pr ( #38320 )
...
Pick https://github.com/apache/doris/pull/38292 ,
https://github.com/apache/doris/pull/34021 ,
https://github.com/apache/doris/pull/38228 , some modify of
https://github.com/apache/doris/pull/37260 , some modify of
https://github.com/apache/doris/pull/37595
2024-07-25 17:32:44 +08:00
54787f51c7
[log](load) print reason when LoadPathMgr failed to delete file ( #38235 ) ( #38340 )
...
backport #38235
2024-07-25 17:28:09 +08:00
a751372e76
[Feature](multi-catalog) Add memory tracker for orc reader/writer and arrow parquet writer。 ( #37257 )
...
## Proposed changes
backport #37234
2024-07-25 13:51:59 +08:00
792bd7c74a
[cherry-pick](branch-2.1) Pick "[Enhancement](group commit)Optimize be select for group commit #35558 " ( #37830 )
...
Pick #35558
2024-07-24 09:21:07 +08:00
ce1c268743
[Fix](load) Fix uninitiated load_type (not a valid value for type) of StreamLoadContext when operate_txn_2pc #38033 ( #38187 )
...
cherry pick from #38033
2024-07-22 17:20:03 +08:00