From 64d024264302723f0d7ce85239333944d1648241 Mon Sep 17 00:00:00 2001 From: Marc-Eric Martel Date: Mon, 6 Nov 2023 13:30:36 -0500 Subject: [PATCH] Manquait un ) --- .../bin/Debug/net7.0/gregorsamsa | Bin 151064 -> 151064 bytes .../bin/Debug/net7.0/gregorsamsa.dll | Bin 9216 -> 9216 bytes .../bin/Debug/net7.0/gregorsamsa.pdb | Bin 11556 -> 11564 bytes .../obj/Debug/net7.0/apphost | Bin 151064 -> 151064 bytes ....GeneratedMSBuildEditorConfig.editorconfig | 2 +- .../obj/Debug/net7.0/gregorsamsa.assets.cache | Bin 6261 -> 6261 bytes ...gregorsamsa.csproj.CoreCompileInputs.cache | 2 +- .../gregorsamsa.csproj.FileListAbsolute.txt | 42 + .../obj/Debug/net7.0/gregorsamsa.dll | Bin 9216 -> 9216 bytes .../net7.0/gregorsamsa.genruntimeconfig.cache | 2 +- .../obj/Debug/net7.0/gregorsamsa.pdb | Bin 11556 -> 11564 bytes .../obj/gregorsamsa.csproj.nuget.dgspec.json | 12 +- .../obj/project.assets.json | 8 +- .../obj/project.nuget.cache | 4 +- dotnet/josefk_producer/Program.cs | 2 +- .../josefk_producer/bin/Debug/net7.0/josefk | Bin 151064 -> 151064 bytes .../bin/Debug/net7.0/josefk.dll | Bin 6656 -> 6656 bytes .../bin/Debug/net7.0/josefk.pdb | Bin 10876 -> 10888 bytes .../josefk_producer/obj/Debug/net7.0/apphost | Bin 151064 -> 151064 bytes ....GeneratedMSBuildEditorConfig.editorconfig | 2 +- .../obj/Debug/net7.0/josefk.assets.cache | Bin 6261 -> 6261 bytes .../josefk.csproj.CoreCompileInputs.cache | 2 +- .../net7.0/josefk.csproj.FileListAbsolute.txt | 42 + .../obj/Debug/net7.0/josefk.dll | Bin 6656 -> 6656 bytes .../net7.0/josefk.genruntimeconfig.cache | 2 +- .../obj/Debug/net7.0/josefk.pdb | Bin 10876 -> 10888 bytes .../obj/josefk.csproj.nuget.dgspec.json | 12 +- .../josefk_producer/obj/project.assets.json | 8 +- .../josefk_producer/obj/project.nuget.cache | 4 +- logs/controller.log | 145 +- logs/kafkaServer-gc.log | 250 +-- logs/log-cleaner.log | 4 +- logs/server.log | 1956 +++++------------ logs/state-change.log | 403 ++-- logs/zookeeper-gc.log | 68 +- 35 files changed, 1027 insertions(+), 1945 deletions(-) diff --git a/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/gregorsamsa b/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/gregorsamsa index b10a4d0d89a83fee29d3e714af1f623bd6592f77..fefc1d5dbe1345c6907183742063805905881c30 100755 GIT binary patch delta 1540 zcma)6ZBUd|6yE3Vvalfr!-t}dqKQBdvg|J3UjwqNfuE=-TC^&ts}7L6IisR1uqgo! zw&`)en!Y4;CS!cbi#ZuWh}u+`DIw*EnSxfN4uo26cb6aj?bm&u=bZaI=e*~<@4-W= zgNIbj5puurw`*Gh|Bh}*><+n*RhVJ>F~BduZe)C6A&VWg-`cW#zsa( zsx{Kw)bXC7nxa^ZF}g?{7gHD&wXP+kF61N7>56h$xu7DY&~|S*toHcM7##Viq?6&Oy|ilrrr4{y4#8cTb^b#yCM%UL zUinpyd-mhK^~je}uQe-i-2T^%MS@E2PeY(o*&nIIfOO){APc>^)aCQ|ro&PB1Qt(G z$jMa>hlG17@pvmwMEivI>r%(vhf4I@`Tay3wm-bT82jw{2Vqzzl{~o2m@-$&9JT?! z*}oZyM3eM*@X;_>87|Vc2m~RFH%6dEA(R%X zp+hTwQ;l@Mk1j+b4Mp^v23l;PR1JcW!V5K!BFeM<$4O#)~#k$OjUz7I>HDB?*Z#+MI+HLa8eWgqu{5 zj8X*BrDT}U&XZCQ>xPBgoPk6jm^aVCA5Mw)^cVsT^SuT@E~=@|#yaev(QMp6Bk#?D zg5hU=eLf6wRP&IcXy6n18)|&Tj+TuDdw5G3CbMx{zj7vxn<*l*dEYJ^299#8 z6?>s6>v-z?)SC2SG&Pobk+J%mi+N{^X?@02IGMWSE=sQyBQemmO00)BhH_4wG`lNsm@=1(s_7! z3nU8MnC5U~+D3N1lvzETQh!=H2j ztVFDmJ7XzGAw6SSrZUhD#!O<*?l6{zOiGorbjGr%QqC6PGrA&Y3s6A83YN<}^|VdF zvd~WL3bq)<Z;Di0PVa_T%-ER_;NjKbN>$(?pq!J delta 1540 zcma)6YfzMB6yE3EdvW`(KM3S-V%jW#bb!k7~kZd9+;Q6Z>6DX3$N5~bFtuG#qDU||=-0c%nBc=%XryZ1pMy?*Y12YXs5 zUAXqE5_hb}dmE81#r3r~OHT>Tto zcTvdKD`i>2A8z9FMyBxg3+;TV^Y@3&=(F98SY_tNuYS)<}g9t}-U4NlA|^ zz)kDL$6+Xuo;}_UT#+nKHbJ0APcHzmQrq(;<)oM-d$&t^Y#vjfFj)GF^D3eh#o0nk zvKGl&oRAL5ohtmsO)9jELIX|m#A`6|I8Rs@?ogK({LxM}FHFT|?&^(Xh74LVPQbso zc^s~fK>;NPVyo--qmrSSN<-lXf4(;qr&I`` zx7CQi3I4Gf34jNk4@W$*=oJmL*g$a_1VG1)8c0gq<`Z2g{gh3SZ=@+c&sk#|LjHn5u8<{%enqLdhsqJ#=!uuNp?jsc;c zvSLw$v2-~WCbaUHxzIUb0WY41XkZF&NyMM_6#MlU0?fR@04PN@jSI03W_q#^1K7=b zlc8d`#&0fyL5V6Jn1(vwAW7-az=!Uo<137(T^XXd2b^bM&>e~V^ED9Ll|buqa2G3R zQ7$C-ab^@XLc#s>z?o3!dl&!U82!8+G5CqRP1uPEbjXBECf^|0gtM-v`d`>ro70c^ zn9F>BA&(6>BmRB&1_ZjGhlbum2d>b`Vx+*!O%~0nhcC=)~*~3qjVGuRwh3mu>*&lSurAMvJ1;(0|{hy zVRzwAD_z+v3{$-;OUGw)%TdD*V>2<#>)lw1J4(6Ahoyu>tHU*6+Hh^S n&M$u6rhlrOK zGcvsv76|joPGn!R)T4FdUS^Ll7sJheL={+g<=?A(U|{qOU;^qbnye`$&w9^-f#Kd{ zUnym!6>gurm|m@Q>|yxibotoV6O2GL`o+mbi6zMy`bDV)`Nflaq=Jp0$}XOPDgzo> z(A+*P;q%4xKY5QDF6aW?1_SOOw!%b3F-FVD8q!*m1EjZY=9Brw$`oU{Sy1Uc<3s~? E0J!u_SO5S3 delta 223 zcmZqhXz-ZO!E$=VqIDB{WEg`ct_BA2MoeyGtYojr?1$|Gg;eow025A5jGs-ZRZ*>lhe)1DJq%b0%v_$+O-78g*l` zuavUP-{9@LFPEkGbZq@QapHoqe~dsi`o+mbi6zMylP5_98^IKufGPkPHbvM-Vbvlv zu4TJ)S=23oI-tN2#8#N7D8^_!Sx;Jv^}i~Eg2QA%rR|&fWPY(Sg;;JDRC>=i(SRKQ DS;9^9 diff --git a/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/gregorsamsa.pdb b/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/gregorsamsa.pdb index 7ede018c3fb7ea71ec622cd4d48dc14f69fb0c96..4c25f4aba5ae693d670f3bd34bfd1432d6c71fc6 100644 GIT binary patch delta 6169 zcmZWt2|SeB`+vulJzKn%LB<-%l0=ai%-9B9gccc$rR+1fl|2<@KT%YcEJFw_qGSmz zN@=+jQCg6#Te;Sf|D0)byTAW@KD;yYe4p>LoaZ^`nNzy6dS_;+VzmZ>;0S>sUI>Cz z5Wm`RjMv~(k*|cOw`cAu{H^>sE0d13Krjdy;XuwYpOwt#0`tj(;XuqW7{n2*#Yk&x zQ^O!6btc8H7Y0$#z#yv3KSQR4pDzqzb=e!Eg3%OKS3@|FFpL_)g`DSXL%0z&^)`eD zF;*X9zMo_M=RwwJAROFC3AoJhoQFG{8X7_aDVzp;yc-Y<-)gL~5xC9iC@8@G6de9% z99YZcC?xpr7xCR7hRfvlZxYnNZ7%Een+&w|HN6bHjr4SNy>z`c>g##?cZ7iHr{T`I-EV55Pa zk%2cwZxd5f=kj~*Xbk!!6750a5u|_D@yJ8b2{ySexEgSw{U)Gpl_W@8C4jcCQprXX zG020hy4AiPux9yw8A!HCmm(7&hZ>H{euLVGnOww!7>aG`EB>l%tK}u|Vbxl6@Q%#i zP}ufPt^Y)aPIHs}N^YcsK(oLJ<3@;fu<`&JLHH%DCV3D7ggDyhFkR z1H80-0s|2Q6aK=q4_fyTQyZ>(Q)YJTR?6+^NsYi@DxF5B1p83Jeb~skndAke4TDto zzUdJu?^?h_YK%-NGwDF;#<{`m!>YsELo`hhA8-_Zed&|q8(nbiywwIOo zrAYNPc(wu|!q=A?PGhqN>AgM-!8Kv0+IQJ`;#>9&wm%s2KmiAzvJ@t=b`i$HkE=}3 z6?C^N^l?+zXHo?)d=m0M!$9V=jxwV`dN}O3?g~7^m$wS;6a`T+B&32#KZG2uJ7F^&{MrqU11(5p$)s!1!BIj@pQ#{zHJf#Rm_*BqXt`9Gu z@Yuwz%}RI%p)ATcIa|MM2T%}@1V;Q~=mxK)9;6LH+6UFXyQKG!&pYup<%lYTu>c47 zB~g(jkGC+cN7@_zxg0TI-XhT3bQ4$23ju83sTCKBA%xt1uM)-GBwu&?IQ}}fJch;b zgx}aHmm?>97Qx_AzLk|-g9(wxD*x5FQ4rJXg54*)u zrVB`*WnXrlK!z~L?3Mg7%P+6(PZG}2$EL6LfFc3exDwyE7tr+D0ri`gT_er2)0|Tq z`xbAomx#Hbw2(k5dx<*Jmrg)a+bVd&o$;QeTEn2R>o07d{sM=uAcMISnjlgNProZC zL$)s0*y1Y)2>aMb3ksd0Z5|R7v^FG|&8=gt(ts3kAbyoH?$b}>$c#Tsp6QX?KY(3= z%7mtBx98By=0~c84zE7jQ+X20iwn^kf2aFd244R4BhavM%B@-DOfG!-cLTy2QvH3< zE{H?xvI>4c8l8?E)1MtG>=F|=`ZYQH|6~fq!D=BvbhNvx1_BNTx6@NIV$AY8cQvSg zT5@Lrh(c26mY%W0$Pj|eW+Q37*8SwntjvXs=TCE3fMg*_loqr9EM&NvuIDv-+t9k` zNOWIA`$7r}a8F1W73a*ZhxPJTU}E%RocIbuxW#^EP zz?Jqg-;;d?I`+m@O~`n6C^;$FRM{lJZySrLMOX|yMAK$%7h9h4H;-)lW}ccSvt5Q) z>!;`6goWe8A~ugfU&LZg1)YiKbYO_*SA7UjlxHu(&-{N4-kl8FaPiK2%v6ED#U*Tj z*vDRRg_V^Gr$hw=`?1$t67BATJ_&8-4jbIlW>MH(FndH|~I{)mSw1`2A0iRrsDI!?CDlc10hU+EGoUOhE;KXs;yBz z#`X?r4sp#dh9**2+IK}nQ31<*mUi+N#dx3NdH5QKp}Sn?SEYe3A~PSY`8S{F@-BzM|5w=Ex0IpHKROsmAfo1FVL1x`s@mZ5D+OUivDnIdI&j& zRlNqMgWRpIgbh|w$G$&<0Q;Q*p~Qe+?R~B{r3*p>I~wF}n7*^ET<>Dj`DOK9c5bX4 z*eMBI*!vH}`EQR=G8H>ds42PrtfFa)Kj0M8ux%Q2{s0~Fh}T7ThwwTiDH}_KZ}(~X zmjqg5fBIQ7Nn6{7A)memW{p-pj|NV^;;fis4W-$G^ToPaa`~#3LS;6ONX~ z3`QlD>VWm)il|WOyHqID^&a+V9*S^ibp1?fwrQRNRB~wIsY;*P`;c z{2fZek$s?0T$y-!$A<|hynSWTFP3kIeO7RdQm&Vp&c6XV288~q)NSdriO|j`rxQM} zllI^bi_!=So0{4HpM)x5z{&(bs^RV5>M!~>kuJolAGqGYZOW#y3kVKlr~F_**BmVP zx}&YZd08`dV7balMc?8Uyg@Ns%tU(>PMXDS?61 zKsKMbwMQ@HE3gi7%IUEqTrW=A?fzu_0O*&HiyJX!4h~GAL1m0bn98{MCC5$KN#-de zHXoT96cT=bP1I~t`5N~7M(#Ti+Yb2>FRELF`c#!2XJhTDKZgMXvwya%4#v7R%(Fi% z@tFOYT1lmmFIFoj0t8yo>7h(oIXIPQG+#Oar;>xqJ1?jWTM|l}z0NmiUZ8-iHL`J^ z#aeH`q~;b_XXkUH7WLY>ANhiG4Sq8Ux$FX%b4f6DrIRi9e@TE^-Lr;hQpX?U{9{g! z&+2{L{hJn<60nk?u~gBQFbSPkj;og15b@_Y2%4Lw3NJCyKvNQn3VW|ff$?m5HU{SN zI?RiRVohr@HQj-yq%!fbz4u9|jG4bapDJx-#Yj#5H2;pP?l)78|89fN9`LjJ>Ek$o z(O2tceWojpgeryvfNn`Sw0b5c2~y$L-qqDx3?!F6uuzuD30n_-NJ^q=F6TyJNF_(B zq$PJ9vdYb#%BbLOy#utQWYE87FAc)8N@zkvr$tH2>`HI+(;puyh`@s>+>j(x0N3qV zLcaa_;3ccH2sBr)`_6ulDJ3758k8^wV|yuEx+{U}l382H7%Og z@h;buR+Yf~u1A$@&K}TAhd!+N&KNH`GeW2eL$E389-aTCbPNyqbEOn*1s_#GjiQC% zdIiM)OZVTXaJ)3u;k&FH zI^q{EPsr4ldzrsyZO)2LL=ZJRzus)B z#NZ5D-R6Orj1+t3{JhSo7*JlrqQIAk+tWGGgd%b>mCW03$<^vGG7S0kjUtRX)zo0v$3qR7UnJ0}Ah_h?u4b z`&n1hUwdZL@9hIWWY(g+ITh~^MGmCsacQO0G~T+fp-biLr==`lD2qcQBitro64Fjv zIdHpQ#aGWrsDGF%c?7~`)zHKXrSp*XQ8t&5pq(@+?Zt$s(4%{!pk7uUt(A#A0)?|1 z1k<8(?phSRkkt=6EtAR=ii1U23AFH47VB_pSBxbyVwM~l7PM|iiX^oIRXKTdD95E1 zYER+j+vv&~4khl6>(&?K%79`yMG!3~gnAXqvNnAhSBm?Yu`T!`{@0e$c28lM0By^T z;3$^uZ)A^;e;v#x=NU+T36UI6URj-pV0u8%%0w6Om)3d2QxFMMmwfv1=jkELR=mB{!dSI-L)3$8(*Vi z=OSHmK>}6%CqcGniY3%VoKb-=O$mR+MG<^rs&^961j^6WJsHg3}xd~+ak2ZR7e1yMBr z_@fe7FAhiDZb)2JA{Hpm^411CNQD5L4eeJyJjNX=hdAU-%WMoVB&P52YLT$%>|+r< zXA<2$IX@0Vc!$HS**WK2w_{4$@f509hQE zDk`HwJ2rPhp{kco8h?5(@i_U9t%kneX2C&4EGn*ka|*`ja>{ECKjme8teYp!IPMuI zs8!TP(G%jYq0M{!n|b%xP|RDmno$vtk6lb5R>}O=knFpSb|1nFtFbPn2_#}g$9Pt@ zs3G ztA_Hwt-D`I)p*;c>X7!{EavlJSB(?ZJ0d)QzOuOondp~8wBrThozYvG7J0EgpEi&t zF8Bcj3Y~g3c($dRv3W>v-d3aThfMfK%24zP*oO1l`_LwzbBF)@?0%lE8of1@^)n?#uoRNsK)4hs&lV2t^g{ksBIwSj+%l*wt?8)Ku*J^&)N#@N zjG7luAz&qM7rxu2!Mv&Z)F+?b^WLnO+@!E!{Gc$4BwbYu&2Dc>fPgVvcU>1n&mrZT zjNU|LNHh!ZkZHw!&ZQ6nYBt@a?mB2-d*Mz}C4NV=DVSB2M``@ze3;m+;&iX3+kaZ8 zcoOm*a$WZVEj1uOD(k56`e#@7`!m-r4keyIQp8YkBAkD>j}v zim_|2-BXiwuVQ!3ZE^=Goyf?x8+)j5?apsN<{%9M)mO(g=-**A;O7r+Enl`CCNy-d zG2BV-UR;3&1p96nq-6Bf>x&NBVo^2cM7k( zyNGgnt>mgTBg^|8qF1U40)PGr!m>8GH;-*u7iF)oBECxvCoEFX4oi+w;e$X*Lmn+) ztQkiT4&<_u{a&Y9Ui(X797t$Qi4!1eDA*co3_l9<>oD0HKX9zUGWCE@@xaJN!fzhA zS%i2qkEmZ0s`!%TDOiNGoVoe$e!YaV+yT^zB$;?{O#_RbsXG}D`4ZKyizaXiAIq#EY*bE0-WJDmO<8-G624?r_M4T=6n< zc>eJ+`&8;!^0euzjf)A9Mv>;{ka+h3TV;HwFjOQ4}D2DOl@;S{TnEwNe CDQ$BA delta 5989 zcmZWt2{@GN+kXeimMzO`8^hSkzK0m=Fvf%yNy```Qw)uzq%1A6MLf1pmXjrwtw_Yt zb~XO7VYqYJ2!gi{4)sS6WWoDt zYfn^wds5>xDi}F!Q%wwF zqs12S>j#5KYhw^a_OC8m!_OZKvi>p+qmEg_r=^MTBD*n~2p^Kq+kvb=l(ae!e#Ag) zg#G_V_FsNPSR3J4ft&`Hd7gp{UQ>aU`Upl)7HbS1@oo?n`uSn^-hBsvHs1yj;a{KQ zZ-PWVo8P`jz=AG5S6xFhJ&KM zE`MBc2!m#vp^YH%2;v)>Ut22W<#O3n`fS<6!~#fIse%sP(v3%O7-X+<-?@cohC`9> zq{7iVQ~hkB4Cq?PkG4NoS%lo8M5SBjGyZmMBlp&70Y5MBiOn5OOgRmS`vzLJc5K_@ zeBnYA_J}uS2hbEyX56+%-XbLka=f_W#=UQE?610)Gm%Yy426W#LRrBKf4zXdB9Vt}*Em;Sp430F@QMMK*79EkZgl$YkMGyPz~HLWb7G zBjfL6BS5VHii!r9zl5f&C}1XAQcDR95B{~CVlABnF9oF07Viz$AhAkDF_kx{+A-~J zhM3s8zwQ7LK{Z!un_3gJbKi;C8{Rk-Iq&H32zrQ~Luha?ofg3eWrlO@ zC={GeWgCtp2-Ls0)V*XEPYi$h`k%8^AX*TQnpmxTi`Ze1z%2*gxpu8_E4>k%eDYvc zG`J$DhOYf8@&u|Xo})(UYAv{*)Oj)XU{tCXObV)_smdQ3ATc2=f$;j!7jnY(?-PBJ zJr_W|kPL7Tl0@r|_Lm_Y2%?a`>BoA$-K5$890BB-e&7I(3dx|x;;kvr%}c;wL`%|101_*=2g?;ClI&92%5|=+6AR^V&kM=QV!wAm`dZB&bte@2 zNbzGwGi7++Z{W~vg(c9p<@*+)ec{P=EnBS{?DLdRvyzFsDjYzpuoOBKlHP%guxlws zHzBtfC8vJd)|buO@dH!~E5s4rtjU4mHTlbJ#|O{4H`fUj9UHRp`4wemcxW)48)2F} zZW^kJcD5yU^N+h$ZxO;0i-`w+fmzGQQfr&95Gg?qo^TvCcPs6d?9nc;ALAk&sS#9m z217zrLYZ9IGp4hzkrEzc*2Q$)7rs#Uw3Gl_H3@nRpo^&4TT-q4zxSwU4DBwH-6=zQl&d62EI;IL0Rz(g;4}b?Ecy!$MR1XwG9=g(rCJ-Et z3zl4Q&YUd({Hv7F^5`9Hkmz|%w!SJY*!9|ok@WZe&lSLCl@8i}zWF&af*>z;){JGn z`{0`F?tzVMy;uk0R%xRv8S!P1lM-{sZdQTfR;%^+c~+Wh59nB>f-Y8nON7n;OtLE^_xqLUf!qgi|o2m|{fJRaExcqq1aVVDe z+^3fxlfM$L4u4-fPg5;g4i|&En;9A&!JvifZDdht4CZcb5|pYv#~|a5FqdabMx&IH zk;>HjvK|Rs#-AZh5Si@?PEU;BfM7D9K%oH?G>Zbmy zAwz_6SxU)fZN&^YB&I^nEIQBxW0HO(=uf$=uUszG?(yI7^r6eLglHx;ghA8uieyGG zLY7O)hZS%c3e+^%G4*1FH7P%wrrr1~_8j;}Odb{0njeIQ&#h3cFFx5#xK0tBMu)AQ z0!48-lx4o@6Xeuiq0oM0)WQ7>&hkj#`of2REUu2;-!Q>xfSkFnYe`AvPDeT?k7WzQ zJOC-;3h2^%+8>ZJYir*6S-md=5Sy`WuHx1!&?c^dCVJ=OK_Y%~FVF7iAiI-(*P8B) zTG)emalE_Gw!90F=$~H`ftJ^kTUC`jup(9bTo+v_u>8YV^zd-@1mi}3*1IYNI$>tp z;{3+Vi%^vQaJ2ei{Z6n)LK7_($en=#O$6Iw%buy)r7z+&zbW+J2InQz?0|JYxeqonKw1{yu2mLfYDA^YJJO0ob21b`2GrgSy0zn01>ZBoYh8L`}@4`{mdW_OR67 zkqn7ZxRFZ}6{X}xC(QGi52l|v3VI|Z&=;!sn-HM;sri_UZwB#%wK^{SqR9^k(AYwa zpas!c+;9+Dhs&WYUTOYT(CEaJtT@9&7C`;F-^Ww?&_(N zSM?_dU9HMKuktyaqh9mRC!dIXN040AyZvI;1USE1&3V^R*JIFvvZC*Q&s)$*m!8(0 zB6j_|Jn`6>*}w>+FCQunlPzDM9>KWhlQY>DovRFQjZzLQYYjlS{p zU}hEh45c7P^7`iu+_1YdnESZC6kT#{gcV;tf6dFWSD(~?jJ+H;TW|+4Qdl%wK79zD zO%7hsW3RV5Q%>9}@T!aYM;6peDa293FxMbwu7zOM>Z03(%)Bd8A5QG?{9T7hj1;RURZ)OC0(hny9~yi-bU8L_`=DU9}GX8SW3_ zMpgv>>Ish|CECB#mfi+XD6N2ww!i0;S+$OqO@{Q}uC4il(z`ut{6VKQHm-X21gDoD zh;}}B7$)LYaL(kA!9HQh-}I2Fj9{*w%&Mvc7@wD|kI!%3;q7{&V9(+6sh=f&Lwis| z=w5X8ITi`m1ukC`HD9Xbu-M`nMsVFQa1!iZo*{VJ?U zzS;(v!25f7orDMct*0P;Lon>$fn$dby(r|O6rZsK}ue^ zr*GSj@EgCW@S^XFq=!c=S8iN+V+%a8ZmR75bL;V z^E1Tk2uFs{S?~<}8Q~qHCHe|gjG()u*``V6+M0(Np(*FJ<&2(z^XtxCJ; z#E*?yi^*HyrCGgSs zw)+2%cVZVAmx(9;W*N-?6jr+3ro7MYjQ5k?jkl9cy`zD9 z-u_s3v zf`(_F7ix)(b!-=3qqNHHn=-hdh*RIh45tPzFBz8o3y+|fh1>5FJp{F#s*W=fYvjA8 z*hoFJ!i)DjBw9$dCF#KV%6%ExRe@pUmPzAJ5`Ehmt>=n&cWs^DJ}LPu`sI8-NZzI|27 zs#B!quZxm5Q6N|2S~oP@?3b0=@l^-~gg-ZkZmlhVO1EDnvf)p(5H zc#&gI2439^&2UU+RPetwHgf>^SVdH7`0*2HT6}p@ofJdIxqVQ%D8Fkc7CgjaQBNzC zXRud1J}AyC^7fNbHU1RXzF!~!*eEN3Fl~PHifYwK$gNgIPVcOINU1h3-B?Zg{Fn{U zL*E<@_y{f5yr)QQnpPoz&Wl?Q&n4(^xJ#6U(XpxNhYA#TEc4Niw(DkEfhg19eAR6@sp)CR|i zD!SU2T*mihK6+LNmT*eoDNX|QDY|zGI$m_-`oB!+RSsFt2Ps*oQ7jvviI$@e(jZY_ z{Z)OGy3X;ETE>TDXR;uhhy{Ksf@r{}bDRa3Zxlnkdu=~4EyJ3WI~+O=#dGW7fa$Jo z*nNHEmWp)2QjfO8&K(18ft4H!`c%Zwzs6n1ppsh?4LS2ZQKTB*V)p&;)I>w$5+B3j1^oq!@E@RGaqvwzs1*n1&y-x0&lKv7K{{hVdQ z(dTe`f8L9?Ld0V8ZtEYr57vOKY6|Gzp6@uO1M<_AgR1fyFOm3J_Z=7tE0!G8t=lgyzroN?Cw%2xp+Iz z@x(YEQ@~WmxpeY6=0oQub}nw}?wBWKE{I`YbSvImRusgdQv?3nbP(o+d?+ia8(}bV zTpdpB)b5?jIy4OW)KyWHsEKc!97Z)Aa4T7IKY{KcT=$bc1H?2mP>-4Dw~#X+s}nQ4 z#nh!b1xG{r=GE9lS#I^dBP?)YdFX!Ho!G)|Vvcfl>h!vvS_lEBH6&2Jx%-^=T7yQd zc-`C1u9>eh_%1)J`@#VXXo#WKeZ|?(f@1=PZ$}G8olkeF%dJi`qDfRg&fKUq z2>LN#G2&2x;A1Q#*Imrglm%o>VN^j^m9xbCt*AsX9MN(}Og?5Ndc=nb!ZelKg7CX$ zp}Af|;9o*B4uUhPHlt2c~E{6VQHnmB4zb55R{ zlK-lfoL$J7ql+@avO-0B+)|n~K6PS?TR@POQe0k=Ag9|T_Y5mbBw7+qGd7b9m3s1) zrTkAzMO_3xM`q(I5y~kMa?%ZIF%dmlzy<#|G`*YKtDxaLQ|@j2m0|8>SFtg=x*wL~ z0D(VmJxZLHh#nvIa!$;5^_CrXf@uzo{DlW5Q2M*(xcno5`+m*V<8@bglW`J~U1=QNEIStm5{lU&2X4oGt z$y=rbHQv3y0&Zw4p?fDAIk)|?O$;NK=|fqdAT!dhd;E9zD(+-TZ6@TX3<%rQ)5DyFgm bQ(1?pypE}Sh^c&mshql8x#l>A2SNT1QNcf1 diff --git a/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/apphost b/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/apphost index b10a4d0d89a83fee29d3e714af1f623bd6592f77..fefc1d5dbe1345c6907183742063805905881c30 100755 GIT binary patch delta 1540 zcma)6ZBUd|6yE3Vvalfr!-t}dqKQBdvg|J3UjwqNfuE=-TC^&ts}7L6IisR1uqgo! zw&`)en!Y4;CS!cbi#ZuWh}u+`DIw*EnSxfN4uo26cb6aj?bm&u=bZaI=e*~<@4-W= zgNIbj5puurw`*Gh|Bh}*><+n*RhVJ>F~BduZe)C6A&VWg-`cW#zsa( zsx{Kw)bXC7nxa^ZF}g?{7gHD&wXP+kF61N7>56h$xu7DY&~|S*toHcM7##Viq?6&Oy|ilrrr4{y4#8cTb^b#yCM%UL zUinpyd-mhK^~je}uQe-i-2T^%MS@E2PeY(o*&nIIfOO){APc>^)aCQ|ro&PB1Qt(G z$jMa>hlG17@pvmwMEivI>r%(vhf4I@`Tay3wm-bT82jw{2Vqzzl{~o2m@-$&9JT?! z*}oZyM3eM*@X;_>87|Vc2m~RFH%6dEA(R%X zp+hTwQ;l@Mk1j+b4Mp^v23l;PR1JcW!V5K!BFeM<$4O#)~#k$OjUz7I>HDB?*Z#+MI+HLa8eWgqu{5 zj8X*BrDT}U&XZCQ>xPBgoPk6jm^aVCA5Mw)^cVsT^SuT@E~=@|#yaev(QMp6Bk#?D zg5hU=eLf6wRP&IcXy6n18)|&Tj+TuDdw5G3CbMx{zj7vxn<*l*dEYJ^299#8 z6?>s6>v-z?)SC2SG&Pobk+J%mi+N{^X?@02IGMWSE=sQyBQemmO00)BhH_4wG`lNsm@=1(s_7! z3nU8MnC5U~+D3N1lvzETQh!=H2j ztVFDmJ7XzGAw6SSrZUhD#!O<*?l6{zOiGorbjGr%QqC6PGrA&Y3s6A83YN<}^|VdF zvd~WL3bq)<Z;Di0PVa_T%-ER_;NjKbN>$(?pq!J delta 1540 zcma)6YfzMB6yE3EdvW`(KM3S-V%jW#bb!k7~kZd9+;Q6Z>6DX3$N5~bFtuG#qDU||=-0c%nBc=%XryZ1pMy?*Y12YXs5 zUAXqE5_hb}dmE81#r3r~OHT>Tto zcTvdKD`i>2A8z9FMyBxg3+;TV^Y@3&=(F98SY_tNuYS)<}g9t}-U4NlA|^ zz)kDL$6+Xuo;}_UT#+nKHbJ0APcHzmQrq(;<)oM-d$&t^Y#vjfFj)GF^D3eh#o0nk zvKGl&oRAL5ohtmsO)9jELIX|m#A`6|I8Rs@?ogK({LxM}FHFT|?&^(Xh74LVPQbso zc^s~fK>;NPVyo--qmrSSN<-lXf4(;qr&I`` zx7CQi3I4Gf34jNk4@W$*=oJmL*g$a_1VG1)8c0gq<`Z2g{gh3SZ=@+c&sk#|LjHn5u8<{%enqLdhsqJ#=!uuNp?jsc;c zvSLw$v2-~WCbaUHxzIUb0WY41XkZF&NyMM_6#MlU0?fR@04PN@jSI03W_q#^1K7=b zlc8d`#&0fyL5V6Jn1(vwAW7-az=!Uo<137(T^XXd2b^bM&>e~V^ED9Ll|buqa2G3R zQ7$C-ab^@XLc#s>z?o3!dl&!U82!8+G5CqRP1uPEbjXBECf^|0gtM-v`d`>ro70c^ zn9F>BA&(6>BmRB&1_ZjGhlbum2d>b`Vx+*!O%~0nhcC=)~*~3qjVGuRwh3mu>*&lSurAMvJ1;(0|{hy zVRzwAD_z+v3{$-;OUGw)%TdD*V>2<#>)lw1J4(6Ahoyu>tHU*6+Hh^S n&M$u6rhv-ppIvt?!u~&33=84W7mz>b3u3%rgI5!38qxJ{vWB FB>>L{5~BbB delta 49 zcmV-10M7sQF!eAfP)kQa3;+NC^hau4H~%1M>5GW0GHCK-lXvOv6$@q$5s;_y-*JKG Hu_#R#{HztB diff --git a/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.CoreCompileInputs.cache b/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.CoreCompileInputs.cache index 130baa4..322787e 100644 --- a/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.CoreCompileInputs.cache +++ b/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.CoreCompileInputs.cache @@ -1 +1 @@ -ae803ef1c48286dbe8e8ba58e1f13618e8c4a416 +2e6ed227385247b3097da8142402f636653d97a5 diff --git a/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.FileListAbsolute.txt b/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.FileListAbsolute.txt index 3d582e1..f6b5b49 100644 --- a/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.FileListAbsolute.txt +++ b/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.FileListAbsolute.txt @@ -124,3 +124,45 @@ /scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.pdb /scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.genruntimeconfig.cache /scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/ref/gregorsamsa.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/gregorsamsa +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/gregorsamsa.deps.json +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/gregorsamsa.runtimeconfig.json +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/gregorsamsa.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/gregorsamsa.pdb +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/Confluent.Kafka.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/linux-arm64/native/librdkafka.so +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/linux-x64/native/alpine-librdkafka.so +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/linux-x64/native/centos6-librdkafka.so +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/linux-x64/native/centos7-librdkafka.so +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/linux-x64/native/librdkafka.so +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/osx-arm64/native/librdkafka.dylib +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/osx-x64/native/librdkafka.dylib +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x64/native/libcrypto-3-x64.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x64/native/libcurl.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x64/native/librdkafka.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x64/native/librdkafkacpp.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x64/native/libssl-3-x64.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x64/native/msvcp140.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x64/native/vcruntime140.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x64/native/zlib1.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x64/native/zstd.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x86/native/libcrypto-3.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x86/native/libcurl.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x86/native/librdkafka.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x86/native/librdkafkacpp.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x86/native/libssl-3.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x86/native/msvcp140.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x86/native/vcruntime140.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x86/native/zlib1.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/bin/Debug/net7.0/runtimes/win-x86/native/zstd.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.AssemblyReference.cache +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.GeneratedMSBuildEditorConfig.editorconfig +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.AssemblyInfoInputs.cache +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.AssemblyInfo.cs +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.CoreCompileInputs.cache +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.csproj.CopyComplete +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/refint/gregorsamsa.dll +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.pdb +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.genruntimeconfig.cache +/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/ref/gregorsamsa.dll diff --git a/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.dll b/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.dll index e1911b238244c3032df1dad89d73d830f1633021..00a972ee39dea7b2e086291cd94db5e958112561 100644 GIT binary patch delta 223 zcmZqhXz-ZO!LsbL$+d|+GK^6ZSBCSkS~4*BGXQ}D149byL`C7rij47$36mQc>lrOK zGcvsv76|joPGn!R)T4FdUS^Ll7sJheL={+g<=?A(U|{qOU;^qbnye`$&w9^-f#Kd{ zUnym!6>gurm|m@Q>|yxibotoV6O2GL`o+mbi6zMy`bDV)`Nflaq=Jp0$}XOPDgzo> z(A+*P;q%4xKY5QDF6aW?1_SOOw!%b3F-FVD8q!*m1EjZY=9Brw$`oU{Sy1Uc<3s~? E0J!u_SO5S3 delta 223 zcmZqhXz-ZO!E$=VqIDB{WEg`ct_BA2MoeyGtYojr?1$|Gg;eow025A5jGs-ZRZ*>lhe)1DJq%b0%v_$+O-78g*l` zuavUP-{9@LFPEkGbZq@QapHoqe~dsi`o+mbi6zMylP5_98^IKufGPkPHbvM-Vbvlv zu4TJ)S=23oI-tN2#8#N7D8^_!Sx;Jv^}i~Eg2QA%rR|&fWPY(Sg;;JDRC>=i(SRKQ DS;9^9 diff --git a/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.genruntimeconfig.cache b/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.genruntimeconfig.cache index dae7078..28a4714 100644 --- a/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.genruntimeconfig.cache +++ b/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.genruntimeconfig.cache @@ -1 +1 @@ -50bc464a59aeb65ac858853faf7456b2e8429d5d +2d8fa65874f17431d718045273de21fcc8a07db0 diff --git a/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.pdb b/dotnet/gregorsamsa_consumer/obj/Debug/net7.0/gregorsamsa.pdb index 7ede018c3fb7ea71ec622cd4d48dc14f69fb0c96..4c25f4aba5ae693d670f3bd34bfd1432d6c71fc6 100644 GIT binary patch delta 6169 zcmZWt2|SeB`+vulJzKn%LB<-%l0=ai%-9B9gccc$rR+1fl|2<@KT%YcEJFw_qGSmz zN@=+jQCg6#Te;Sf|D0)byTAW@KD;yYe4p>LoaZ^`nNzy6dS_;+VzmZ>;0S>sUI>Cz z5Wm`RjMv~(k*|cOw`cAu{H^>sE0d13Krjdy;XuwYpOwt#0`tj(;XuqW7{n2*#Yk&x zQ^O!6btc8H7Y0$#z#yv3KSQR4pDzqzb=e!Eg3%OKS3@|FFpL_)g`DSXL%0z&^)`eD zF;*X9zMo_M=RwwJAROFC3AoJhoQFG{8X7_aDVzp;yc-Y<-)gL~5xC9iC@8@G6de9% z99YZcC?xpr7xCR7hRfvlZxYnNZ7%Een+&w|HN6bHjr4SNy>z`c>g##?cZ7iHr{T`I-EV55Pa zk%2cwZxd5f=kj~*Xbk!!6750a5u|_D@yJ8b2{ySexEgSw{U)Gpl_W@8C4jcCQprXX zG020hy4AiPux9yw8A!HCmm(7&hZ>H{euLVGnOww!7>aG`EB>l%tK}u|Vbxl6@Q%#i zP}ufPt^Y)aPIHs}N^YcsK(oLJ<3@;fu<`&JLHH%DCV3D7ggDyhFkR z1H80-0s|2Q6aK=q4_fyTQyZ>(Q)YJTR?6+^NsYi@DxF5B1p83Jeb~skndAke4TDto zzUdJu?^?h_YK%-NGwDF;#<{`m!>YsELo`hhA8-_Zed&|q8(nbiywwIOo zrAYNPc(wu|!q=A?PGhqN>AgM-!8Kv0+IQJ`;#>9&wm%s2KmiAzvJ@t=b`i$HkE=}3 z6?C^N^l?+zXHo?)d=m0M!$9V=jxwV`dN}O3?g~7^m$wS;6a`T+B&32#KZG2uJ7F^&{MrqU11(5p$)s!1!BIj@pQ#{zHJf#Rm_*BqXt`9Gu z@Yuwz%}RI%p)ATcIa|MM2T%}@1V;Q~=mxK)9;6LH+6UFXyQKG!&pYup<%lYTu>c47 zB~g(jkGC+cN7@_zxg0TI-XhT3bQ4$23ju83sTCKBA%xt1uM)-GBwu&?IQ}}fJch;b zgx}aHmm?>97Qx_AzLk|-g9(wxD*x5FQ4rJXg54*)u zrVB`*WnXrlK!z~L?3Mg7%P+6(PZG}2$EL6LfFc3exDwyE7tr+D0ri`gT_er2)0|Tq z`xbAomx#Hbw2(k5dx<*Jmrg)a+bVd&o$;QeTEn2R>o07d{sM=uAcMISnjlgNProZC zL$)s0*y1Y)2>aMb3ksd0Z5|R7v^FG|&8=gt(ts3kAbyoH?$b}>$c#Tsp6QX?KY(3= z%7mtBx98By=0~c84zE7jQ+X20iwn^kf2aFd244R4BhavM%B@-DOfG!-cLTy2QvH3< zE{H?xvI>4c8l8?E)1MtG>=F|=`ZYQH|6~fq!D=BvbhNvx1_BNTx6@NIV$AY8cQvSg zT5@Lrh(c26mY%W0$Pj|eW+Q37*8SwntjvXs=TCE3fMg*_loqr9EM&NvuIDv-+t9k` zNOWIA`$7r}a8F1W73a*ZhxPJTU}E%RocIbuxW#^EP zz?Jqg-;;d?I`+m@O~`n6C^;$FRM{lJZySrLMOX|yMAK$%7h9h4H;-)lW}ccSvt5Q) z>!;`6goWe8A~ugfU&LZg1)YiKbYO_*SA7UjlxHu(&-{N4-kl8FaPiK2%v6ED#U*Tj z*vDRRg_V^Gr$hw=`?1$t67BATJ_&8-4jbIlW>MH(FndH|~I{)mSw1`2A0iRrsDI!?CDlc10hU+EGoUOhE;KXs;yBz z#`X?r4sp#dh9**2+IK}nQ31<*mUi+N#dx3NdH5QKp}Sn?SEYe3A~PSY`8S{F@-BzM|5w=Ex0IpHKROsmAfo1FVL1x`s@mZ5D+OUivDnIdI&j& zRlNqMgWRpIgbh|w$G$&<0Q;Q*p~Qe+?R~B{r3*p>I~wF}n7*^ET<>Dj`DOK9c5bX4 z*eMBI*!vH}`EQR=G8H>ds42PrtfFa)Kj0M8ux%Q2{s0~Fh}T7ThwwTiDH}_KZ}(~X zmjqg5fBIQ7Nn6{7A)memW{p-pj|NV^;;fis4W-$G^ToPaa`~#3LS;6ONX~ z3`QlD>VWm)il|WOyHqID^&a+V9*S^ibp1?fwrQRNRB~wIsY;*P`;c z{2fZek$s?0T$y-!$A<|hynSWTFP3kIeO7RdQm&Vp&c6XV288~q)NSdriO|j`rxQM} zllI^bi_!=So0{4HpM)x5z{&(bs^RV5>M!~>kuJolAGqGYZOW#y3kVKlr~F_**BmVP zx}&YZd08`dV7balMc?8Uyg@Ns%tU(>PMXDS?61 zKsKMbwMQ@HE3gi7%IUEqTrW=A?fzu_0O*&HiyJX!4h~GAL1m0bn98{MCC5$KN#-de zHXoT96cT=bP1I~t`5N~7M(#Ti+Yb2>FRELF`c#!2XJhTDKZgMXvwya%4#v7R%(Fi% z@tFOYT1lmmFIFoj0t8yo>7h(oIXIPQG+#Oar;>xqJ1?jWTM|l}z0NmiUZ8-iHL`J^ z#aeH`q~;b_XXkUH7WLY>ANhiG4Sq8Ux$FX%b4f6DrIRi9e@TE^-Lr;hQpX?U{9{g! z&+2{L{hJn<60nk?u~gBQFbSPkj;og15b@_Y2%4Lw3NJCyKvNQn3VW|ff$?m5HU{SN zI?RiRVohr@HQj-yq%!fbz4u9|jG4bapDJx-#Yj#5H2;pP?l)78|89fN9`LjJ>Ek$o z(O2tceWojpgeryvfNn`Sw0b5c2~y$L-qqDx3?!F6uuzuD30n_-NJ^q=F6TyJNF_(B zq$PJ9vdYb#%BbLOy#utQWYE87FAc)8N@zkvr$tH2>`HI+(;puyh`@s>+>j(x0N3qV zLcaa_;3ccH2sBr)`_6ulDJ3758k8^wV|yuEx+{U}l382H7%Og z@h;buR+Yf~u1A$@&K}TAhd!+N&KNH`GeW2eL$E389-aTCbPNyqbEOn*1s_#GjiQC% zdIiM)OZVTXaJ)3u;k&FH zI^q{EPsr4ldzrsyZO)2LL=ZJRzus)B z#NZ5D-R6Orj1+t3{JhSo7*JlrqQIAk+tWGGgd%b>mCW03$<^vGG7S0kjUtRX)zo0v$3qR7UnJ0}Ah_h?u4b z`&n1hUwdZL@9hIWWY(g+ITh~^MGmCsacQO0G~T+fp-biLr==`lD2qcQBitro64Fjv zIdHpQ#aGWrsDGF%c?7~`)zHKXrSp*XQ8t&5pq(@+?Zt$s(4%{!pk7uUt(A#A0)?|1 z1k<8(?phSRkkt=6EtAR=ii1U23AFH47VB_pSBxbyVwM~l7PM|iiX^oIRXKTdD95E1 zYER+j+vv&~4khl6>(&?K%79`yMG!3~gnAXqvNnAhSBm?Yu`T!`{@0e$c28lM0By^T z;3$^uZ)A^;e;v#x=NU+T36UI6URj-pV0u8%%0w6Om)3d2QxFMMmwfv1=jkELR=mB{!dSI-L)3$8(*Vi z=OSHmK>}6%CqcGniY3%VoKb-=O$mR+MG<^rs&^961j^6WJsHg3}xd~+ak2ZR7e1yMBr z_@fe7FAhiDZb)2JA{Hpm^411CNQD5L4eeJyJjNX=hdAU-%WMoVB&P52YLT$%>|+r< zXA<2$IX@0Vc!$HS**WK2w_{4$@f509hQE zDk`HwJ2rPhp{kco8h?5(@i_U9t%kneX2C&4EGn*ka|*`ja>{ECKjme8teYp!IPMuI zs8!TP(G%jYq0M{!n|b%xP|RDmno$vtk6lb5R>}O=knFpSb|1nFtFbPn2_#}g$9Pt@ zs3G ztA_Hwt-D`I)p*;c>X7!{EavlJSB(?ZJ0d)QzOuOondp~8wBrThozYvG7J0EgpEi&t zF8Bcj3Y~g3c($dRv3W>v-d3aThfMfK%24zP*oO1l`_LwzbBF)@?0%lE8of1@^)n?#uoRNsK)4hs&lV2t^g{ksBIwSj+%l*wt?8)Ku*J^&)N#@N zjG7luAz&qM7rxu2!Mv&Z)F+?b^WLnO+@!E!{Gc$4BwbYu&2Dc>fPgVvcU>1n&mrZT zjNU|LNHh!ZkZHw!&ZQ6nYBt@a?mB2-d*Mz}C4NV=DVSB2M``@ze3;m+;&iX3+kaZ8 zcoOm*a$WZVEj1uOD(k56`e#@7`!m-r4keyIQp8YkBAkD>j}v zim_|2-BXiwuVQ!3ZE^=Goyf?x8+)j5?apsN<{%9M)mO(g=-**A;O7r+Enl`CCNy-d zG2BV-UR;3&1p96nq-6Bf>x&NBVo^2cM7k( zyNGgnt>mgTBg^|8qF1U40)PGr!m>8GH;-*u7iF)oBECxvCoEFX4oi+w;e$X*Lmn+) ztQkiT4&<_u{a&Y9Ui(X797t$Qi4!1eDA*co3_l9<>oD0HKX9zUGWCE@@xaJN!fzhA zS%i2qkEmZ0s`!%TDOiNGoVoe$e!YaV+yT^zB$;?{O#_RbsXG}D`4ZKyizaXiAIq#EY*bE0-WJDmO<8-G624?r_M4T=6n< zc>eJ+`&8;!^0euzjf)A9Mv>;{ka+h3TV;HwFjOQ4}D2DOl@;S{TnEwNe CDQ$BA delta 5989 zcmZWt2{@GN+kXeimMzO`8^hSkzK0m=Fvf%yNy```Qw)uzq%1A6MLf1pmXjrwtw_Yt zb~XO7VYqYJ2!gi{4)sS6WWoDt zYfn^wds5>xDi}F!Q%wwF zqs12S>j#5KYhw^a_OC8m!_OZKvi>p+qmEg_r=^MTBD*n~2p^Kq+kvb=l(ae!e#Ag) zg#G_V_FsNPSR3J4ft&`Hd7gp{UQ>aU`Upl)7HbS1@oo?n`uSn^-hBsvHs1yj;a{KQ zZ-PWVo8P`jz=AG5S6xFhJ&KM zE`MBc2!m#vp^YH%2;v)>Ut22W<#O3n`fS<6!~#fIse%sP(v3%O7-X+<-?@cohC`9> zq{7iVQ~hkB4Cq?PkG4NoS%lo8M5SBjGyZmMBlp&70Y5MBiOn5OOgRmS`vzLJc5K_@ zeBnYA_J}uS2hbEyX56+%-XbLka=f_W#=UQE?610)Gm%Yy426W#LRrBKf4zXdB9Vt}*Em;Sp430F@QMMK*79EkZgl$YkMGyPz~HLWb7G zBjfL6BS5VHii!r9zl5f&C}1XAQcDR95B{~CVlABnF9oF07Viz$AhAkDF_kx{+A-~J zhM3s8zwQ7LK{Z!un_3gJbKi;C8{Rk-Iq&H32zrQ~Luha?ofg3eWrlO@ zC={GeWgCtp2-Ls0)V*XEPYi$h`k%8^AX*TQnpmxTi`Ze1z%2*gxpu8_E4>k%eDYvc zG`J$DhOYf8@&u|Xo})(UYAv{*)Oj)XU{tCXObV)_smdQ3ATc2=f$;j!7jnY(?-PBJ zJr_W|kPL7Tl0@r|_Lm_Y2%?a`>BoA$-K5$890BB-e&7I(3dx|x;;kvr%}c;wL`%|101_*=2g?;ClI&92%5|=+6AR^V&kM=QV!wAm`dZB&bte@2 zNbzGwGi7++Z{W~vg(c9p<@*+)ec{P=EnBS{?DLdRvyzFsDjYzpuoOBKlHP%guxlws zHzBtfC8vJd)|buO@dH!~E5s4rtjU4mHTlbJ#|O{4H`fUj9UHRp`4wemcxW)48)2F} zZW^kJcD5yU^N+h$ZxO;0i-`w+fmzGQQfr&95Gg?qo^TvCcPs6d?9nc;ALAk&sS#9m z217zrLYZ9IGp4hzkrEzc*2Q$)7rs#Uw3Gl_H3@nRpo^&4TT-q4zxSwU4DBwH-6=zQl&d62EI;IL0Rz(g;4}b?Ecy!$MR1XwG9=g(rCJ-Et z3zl4Q&YUd({Hv7F^5`9Hkmz|%w!SJY*!9|ok@WZe&lSLCl@8i}zWF&af*>z;){JGn z`{0`F?tzVMy;uk0R%xRv8S!P1lM-{sZdQTfR;%^+c~+Wh59nB>f-Y8nON7n;OtLE^_xqLUf!qgi|o2m|{fJRaExcqq1aVVDe z+^3fxlfM$L4u4-fPg5;g4i|&En;9A&!JvifZDdht4CZcb5|pYv#~|a5FqdabMx&IH zk;>HjvK|Rs#-AZh5Si@?PEU;BfM7D9K%oH?G>Zbmy zAwz_6SxU)fZN&^YB&I^nEIQBxW0HO(=uf$=uUszG?(yI7^r6eLglHx;ghA8uieyGG zLY7O)hZS%c3e+^%G4*1FH7P%wrrr1~_8j;}Odb{0njeIQ&#h3cFFx5#xK0tBMu)AQ z0!48-lx4o@6Xeuiq0oM0)WQ7>&hkj#`of2REUu2;-!Q>xfSkFnYe`AvPDeT?k7WzQ zJOC-;3h2^%+8>ZJYir*6S-md=5Sy`WuHx1!&?c^dCVJ=OK_Y%~FVF7iAiI-(*P8B) zTG)emalE_Gw!90F=$~H`ftJ^kTUC`jup(9bTo+v_u>8YV^zd-@1mi}3*1IYNI$>tp z;{3+Vi%^vQaJ2ei{Z6n)LK7_($en=#O$6Iw%buy)r7z+&zbW+J2InQz?0|JYxeqonKw1{yu2mLfYDA^YJJO0ob21b`2GrgSy0zn01>ZBoYh8L`}@4`{mdW_OR67 zkqn7ZxRFZ}6{X}xC(QGi52l|v3VI|Z&=;!sn-HM;sri_UZwB#%wK^{SqR9^k(AYwa zpas!c+;9+Dhs&WYUTOYT(CEaJtT@9&7C`;F-^Ww?&_(N zSM?_dU9HMKuktyaqh9mRC!dIXN040AyZvI;1USE1&3V^R*JIFvvZC*Q&s)$*m!8(0 zB6j_|Jn`6>*}w>+FCQunlPzDM9>KWhlQY>DovRFQjZzLQYYjlS{p zU}hEh45c7P^7`iu+_1YdnESZC6kT#{gcV;tf6dFWSD(~?jJ+H;TW|+4Qdl%wK79zD zO%7hsW3RV5Q%>9}@T!aYM;6peDa293FxMbwu7zOM>Z03(%)Bd8A5QG?{9T7hj1;RURZ)OC0(hny9~yi-bU8L_`=DU9}GX8SW3_ zMpgv>>Ish|CECB#mfi+XD6N2ww!i0;S+$OqO@{Q}uC4il(z`ut{6VKQHm-X21gDoD zh;}}B7$)LYaL(kA!9HQh-}I2Fj9{*w%&Mvc7@wD|kI!%3;q7{&V9(+6sh=f&Lwis| z=w5X8ITi`m1ukC`HD9Xbu-M`nMsVFQa1!iZo*{VJ?U zzS;(v!25f7orDMct*0P;Lon>$fn$dby(r|O6rZsK}ue^ zr*GSj@EgCW@S^XFq=!c=S8iN+V+%a8ZmR75bL;V z^E1Tk2uFs{S?~<}8Q~qHCHe|gjG()u*``V6+M0(Np(*FJ<&2(z^XtxCJ; z#E*?yi^*HyrCGgSs zw)+2%cVZVAmx(9;W*N-?6jr+3ro7MYjQ5k?jkl9cy`zD9 z-u_s3v zf`(_F7ix)(b!-=3qqNHHn=-hdh*RIh45tPzFBz8o3y+|fh1>5FJp{F#s*W=fYvjA8 z*hoFJ!i)DjBw9$dCF#KV%6%ExRe@pUmPzAJ5`Ehmt>=n&cWs^DJ}LPu`sI8-NZzI|27 zs#B!quZxm5Q6N|2S~oP@?3b0=@l^-~gg-ZkZmlhVO1EDnvf)p(5H zc#&gI2439^&2UU+RPetwHgf>^SVdH7`0*2HT6}p@ofJdIxqVQ%D8Fkc7CgjaQBNzC zXRud1J}AyC^7fNbHU1RXzF!~!*eEN3Fl~PHifYwK$gNgIPVcOINU1h3-B?Zg{Fn{U zL*E<@_y{f5yr)QQnpPoz&Wl?Q&n4(^xJ#6U(XpxNhYA#TEc4Niw(DkEfhg19eAR6@sp)CR|i zD!SU2T*mihK6+LNmT*eoDNX|QDY|zGI$m_-`oB!+RSsFt2Ps*oQ7jvviI$@e(jZY_ z{Z)OGy3X;ETE>TDXR;uhhy{Ksf@r{}bDRa3Zxlnkdu=~4EyJ3WI~+O=#dGW7fa$Jo z*nNHEmWp)2QjfO8&K(18ft4H!`c%Zwzs6n1ppsh?4LS2ZQKTB*V)p&;)I>w$5+B3j1^oq!@E@RGaqvwzs1*n1&y-x0&lKv7K{{hVdQ z(dTe`f8L9?Ld0V8ZtEYr57vOKY6|Gzp6@uO1M<_AgR1fyFOm3J_Z=7tE0!G8t=lgyzroN?Cw%2xp+Iz z@x(YEQ@~WmxpeY6=0oQub}nw}?wBWKE{I`YbSvImRusgdQv?3nbP(o+d?+ia8(}bV zTpdpB)b5?jIy4OW)KyWHsEKc!97Z)Aa4T7IKY{KcT=$bc1H?2mP>-4Dw~#X+s}nQ4 z#nh!b1xG{r=GE9lS#I^dBP?)YdFX!Ho!G)|Vvcfl>h!vvS_lEBH6&2Jx%-^=T7yQd zc-`C1u9>eh_%1)J`@#VXXo#WKeZ|?(f@1=PZ$}G8olkeF%dJi`qDfRg&fKUq z2>LN#G2&2x;A1Q#*Imrglm%o>VN^j^m9xbCt*AsX9MN(}Og?5Ndc=nb!ZelKg7CX$ zp}Af|;9o*B4uUhPHlt2c~E{6VQHnmB4zb55R{ zlK-lfoL$J7ql+@avO-0B+)|n~K6PS?TR@POQe0k=Ag9|T_Y5mbBw7+qGd7b9m3s1) zrTkAzMO_3xM`q(I5y~kMa?%ZIF%dmlzy<#|G`*YKtDxaLQ|@j2m0|8>SFtg=x*wL~ z0D(VmJxZLHh#nvIa!$;5^_CrXf@uzo{DlW5Q2M*(xcno5`+m*V<8@bglW`J~U1=QNEIStm5{lU&2X4oGt z$y=rbHQv3y0&Zw4p?fDAIk)|?O$;NK=|fqdAT!dhd;E9zD(+-TZ6@TX3<%rQ)5DyFgm bQ(1?pypE}Sh^c&mshql8x#l>A2SNT1QNcf1 diff --git a/dotnet/gregorsamsa_consumer/obj/gregorsamsa.csproj.nuget.dgspec.json b/dotnet/gregorsamsa_consumer/obj/gregorsamsa.csproj.nuget.dgspec.json index 214a40e..7c565b7 100644 --- a/dotnet/gregorsamsa_consumer/obj/gregorsamsa.csproj.nuget.dgspec.json +++ b/dotnet/gregorsamsa_consumer/obj/gregorsamsa.csproj.nuget.dgspec.json @@ -1,17 +1,17 @@ { "format": 1, "restore": { - "/scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj": {} + "/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj": {} }, "projects": { - "/scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj": { + "/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj": { "version": "1.0.0", "restore": { - "projectUniqueName": "/scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj", + "projectUniqueName": "/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj", "projectName": "gregorsamsa", - "projectPath": "/scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj", + "projectPath": "/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj", "packagesPath": "/home/memartel/.nuget/packages/", - "outputPath": "/scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/", + "outputPath": "/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/", "projectStyle": "PackageReference", "configFilePaths": [ "/home/memartel/.nuget/NuGet/NuGet.Config" @@ -59,7 +59,7 @@ "privateAssets": "all" } }, - "runtimeIdentifierGraphPath": "/opt/dotnet-sdk-bin-7.0/sdk/7.0.401/RuntimeIdentifierGraph.json" + "runtimeIdentifierGraphPath": "/opt/dotnet-sdk-bin-7.0/sdk/7.0.403/RuntimeIdentifierGraph.json" } } } diff --git a/dotnet/gregorsamsa_consumer/obj/project.assets.json b/dotnet/gregorsamsa_consumer/obj/project.assets.json index dcdd14a..ab8767e 100644 --- a/dotnet/gregorsamsa_consumer/obj/project.assets.json +++ b/dotnet/gregorsamsa_consumer/obj/project.assets.json @@ -257,11 +257,11 @@ "project": { "version": "1.0.0", "restore": { - "projectUniqueName": "/scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj", + "projectUniqueName": "/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj", "projectName": "gregorsamsa", - "projectPath": "/scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj", + "projectPath": "/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj", "packagesPath": "/home/memartel/.nuget/packages/", - "outputPath": "/scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/", + "outputPath": "/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/obj/", "projectStyle": "PackageReference", "configFilePaths": [ "/home/memartel/.nuget/NuGet/NuGet.Config" @@ -309,7 +309,7 @@ "privateAssets": "all" } }, - "runtimeIdentifierGraphPath": "/opt/dotnet-sdk-bin-7.0/sdk/7.0.401/RuntimeIdentifierGraph.json" + "runtimeIdentifierGraphPath": "/opt/dotnet-sdk-bin-7.0/sdk/7.0.403/RuntimeIdentifierGraph.json" } } } diff --git a/dotnet/gregorsamsa_consumer/obj/project.nuget.cache b/dotnet/gregorsamsa_consumer/obj/project.nuget.cache index f722e4a..7ba2731 100644 --- a/dotnet/gregorsamsa_consumer/obj/project.nuget.cache +++ b/dotnet/gregorsamsa_consumer/obj/project.nuget.cache @@ -1,8 +1,8 @@ { "version": 2, - "dgSpecHash": "hbkwrQ8l06yJ4wDtZ8S/lhzqbXrFDTa1QFx5WBMg5yR/H0jCMW4oe1E25sZE2kATBFfgwG0oEffTVplzqRur+Q==", + "dgSpecHash": "luilSaEAAkVU0Zu/GS3DtbnENafSKB4Zsn6Fuk5fb7oHgwHvcprvBlVjVCKLX0HLy3EIQiovRrisED21Z4MdSw==", "success": true, - "projectFilePath": "/scratch/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj", + "projectFilePath": "/scratch/repos/Exemple_Kafka/dotnet/gregorsamsa_consumer/gregorsamsa.csproj", "expectedPackageFiles": [ "/home/memartel/.nuget/packages/confluent.kafka/2.3.0/confluent.kafka.2.3.0.nupkg.sha512", "/home/memartel/.nuget/packages/librdkafka.redist/2.3.0/librdkafka.redist.2.3.0.nupkg.sha512", diff --git a/dotnet/josefk_producer/Program.cs b/dotnet/josefk_producer/Program.cs index 585bde5..4486f27 100644 --- a/dotnet/josefk_producer/Program.cs +++ b/dotnet/josefk_producer/Program.cs @@ -13,7 +13,7 @@ class Program BootstrapServers = "localhost:9092" }; using var producer = new ProducerBuilder(config).Build(); - Console.Write("Enter message (type 'quit' to quit: "); + Console.Write("Enter message (type 'quit' to quit): "); mess = Console.ReadLine(); var topic = "test-topic"; var message = new Message { Value = mess }; diff --git a/dotnet/josefk_producer/bin/Debug/net7.0/josefk b/dotnet/josefk_producer/bin/Debug/net7.0/josefk index 195002cd9b8fc9a3da40bb4b0af564ce56029d69..0913ffe4c684f5c988592c6accd351a9feca9312 100755 GIT binary patch delta 1540 zcma)6ZBUd|6yE3Vvalfr!-t}dqKQBdvg|J3UjwqNfuE=-TC^&ts}7L6IisR1uqgo! zw&`)en!Y4;CS!cbi#ZuWh}u+`DIw*EnSxfN4uo26cb6aj?bm&u=bZaI=e*~<@4-W= zgNIbj5puurw`*Gh|Bh}*><+n*RhVJ>F~BduZe)C6A&VWg-`cW#zsa( zsx{Kw)bXC7nxa^ZF}g?{7gHD&wXP+kF61N7>56h$xu7DY&~|S*toHcM7##Viq?6&Oy|ilrrr4{y4#8cTb^b#yCM%UL zUinpyd-mhK^~je}uQe-i-2T^%MS@E2PeY(o*&nIIfOO){APc>^)aCQ|ro&PB1Qt(G z$jMa>hlG17@pvmwMEivI>r%(vhf4I@`Tay3wm-bT82jw{2Vqzzl{~o2m@-$&9JT?! z*}oZyM3eM*@X;_>87|Vc2m~RFH%6dEA(R%X zp+hTwQ;l@Mk1j+b4Mp^v23l;PR1JcW!V5K!BFeM<$4O#)~#k$OjUz7I>HDB?*Z#+MI+HLa8eWgqu{5 zj8X*BrDT}U&XZCQ>xPBgoPk6jm^aVCA5Mw)^cVsT^SuT@E~=@|#yaev(QMp6Bk#?D zg5hU=eLf6wRP&IcXy6n18)|&Tj+TuDdw5G3CbMx{zj7vxn<*l*dEYJ^299#8 z6?>s6>v-z?)SC2SG&Pobk+J%mi+N{^X?@02IGMWSE=sQyBQemmO00)BhH_4wG`lNsm@=1(s_7! z3nU8MnC5U~+D3N1lvzETQh!=H2j ztVFDmJ7XzGAw6SSrZUhD#!O<*?l6{zOiGorbjGr%QqC6PGrA&Y3s6A83YN<}^|VdF zvd~WL3bq)<Z;Di0PVa_T%-ER_;NjKbN>$(?pq!J delta 1540 zcma)6YfzMB6yE3EdvW`(KM3S-V%jW#bb!k7~kZd9+;Q6Z>6DX3$N5~bFtuG#qDU||=-0c%nBc=%XryZ1pMy?*Y12YXs5 zUAXqE5_hb}dmE81#r3r~OHT>Tto zcTvdKD`i>2A8z9FMyBxg3+;TV^Y@3&=(F98SY_tNuYS)<}g9t}-U4NlA|^ zz)kDL$6+Xuo;}_UT#+nKHbJ0APcHzmQrq(;<)oM-d$&t^Y#vjfFj)GF^D3eh#o0nk zvKGl&oRAL5ohtmsO)9jELIX|m#A`6|I8Rs@?ogK({LxM}FHFT|?&^(Xh74LVPQbso zc^s~fK>;NPVyo--qmrSSN<-lXf4(;qr&I`` zx7CQi3I4Gf34jNk4@W$*=oJmL*g$a_1VG1)8c0gq<`Z2g{gh3SZ=@+c&sk#|LjHn5u8<{%enqLdhsqJ#=!uuNp?jsc;c zvSLw$v2-~WCbaUHxzIUb0WY41XkZF&NyMM_6#MlU0?fR@04PN@jSI03W_q#^1K7=b zlc8d`#&0fyL5V6Jn1(vwAW7-az=!Uo<137(T^XXd2b^bM&>e~V^ED9Ll|buqa2G3R zQ7$C-ab^@XLc#s>z?o3!dl&!U82!8+G5CqRP1uPEbjXBECf^|0gtM-v`d`>ro70c^ zn9F>BA&(6>BmRB&1_ZjGhlbum2d>b`Vx+*!O%~0nhcC=)~*~3qjVGuRwh3mu>*&lSurAMvJ1;(0|{hy zVRzwAD_z+v3{$-;OUGw)%TdD*V>2<#>)lw1J4(6Ahoyu>tHU*6+Hh^S n&M$u6rhtxdVU%snDq9WMcQnzhh z!V|+}_=3xH7Xzbj029#O3?QxmdZ&UP$bbSyAT12U>OlMeDEDBpA-}THLYB=>N~6#E zHcn!^P;{hiI!Kj%adJ^&NpgmMQEEYc@#F%2Pa~)@-h)tOAj@O}v|hjO*Q)nC|FlgX q7^pDd31TZuR1^b-9+QC9W<7x|jGOs{ez7ve*liY+c+WV|fE@rTm1Q^p delta 353 zcmZoLX)u}4!IBuh@$keR8AiQ{E5rG2>M}6+GXQ}D1A`6gL`7kq1O^5kApegJ14GDU zMJAoek&Nn$CX*W(>lt|`GcpPDL~t`OFalMu0ny|{0fWf_Omi6>C%xT`3v#S}adJ^&Npi;IdHkM6PzBYTPz4}E z|9?7mXzuiLYpgyR^)u=LbwGhMh^;VDQ4HvN0|6~|28REt3<_+=9$L5-QWE4e&#*j^F8M|=RD6j@3Z8R>Y5!UZ>^3XN`#4 z*)PA3^6Voe$o-`81A(=WdnJBuO&yHE$8oEwBX+{bD`g17cH66!yclc@s=E!x-F9(oX4)?<7L6A z!L);55bMOx_wS3u4xrDr>GwBDI1s~TvwowYw!Ws1p|7!?uC9-+kAc3Pub;1uzs@GI zKgCy1*GON_*hoj$2maUf1v6}M>-07C{f+d=dj7hbHd1sp`5NmOl6}Yq2D%iop`n4H zp0S}XS#KjlQ|IzY_E-%1I2!Fk5)tHu=ke%6u}QZ1gX~R8VJ}PpnNu8W<>W;>*Qn$o z@)+d77TsF^_c*h{fGi}}v|FA5kU@ojdVVk*o%V8Ym`w;bp3a@79OSQ zBwaoa`(i(LxqYTJ-Qe>YNfepaItr_;4{OhUm-i{YvYpS1@4LkckmpiDKekvuLCi2n zjBUdjckdayT#uRECg*wMfhU(TN}H7|g0v?nTRAYMGLGjkr6bMCGxDH-O9hQ*e?JI? z$EG|tDdOpbiWrmBT>XljV1Qd3OmYdM8+?}gkPZat98p_x&Fmwccjjuz6Os#O0@iVh zqk_wO-@qk3)HeA0a@4SS8}C5NO{H2cCLn~H6QxvLD1`u0=e=rF=_cvA`^U*w`IT`@ zz%BHAQdyFAhL({(}FG2s3n)`6_#ILIh-V%i=3Ff+6T7tNGFu}Ck#T<3d8C*FS|vX=VrL1 zHxGTk!O9IIleEwv3M)5Vnad}jscltU5iWRdV!cuD#PvbDC;xyWR*@m>axD-kfsec^ zBSo^Q)Y$CL2MGIENDFc#S=&4`I2cUvaHBdlDoscU773`9@|b;+K%(z3eX2+Fcn|n^ zl?g4i?$4lutq)ZRw|$1}3Qmizd|!+;_?_uj1^Waxet#M`$SL{>X>@Myoc-chWuKgceW8gB`~%eS%Aw=E-E~lSIHWT&JuA+vu*)De$pvtF#(&F}?g_>70^?c^;7}=B@i5+U{ zTucK2dDO0}fZF3`S(IQ9Y+emT^9Pwp9L^R=5Azsc?}QcCJp zWm9UvHn5vt1wBO5X72XQPk35CZ2MuJep+g~6qi;SIL)t+ARM)60@@UgD+oT5$mU2F zDXe)dxD3fY|sx6${WcbJ)?K#NN_s_@4F5xLcXN04KvApxw45J!85pfCK} zIl@QwbXXMknHszqPM7>Yuu}vzm>LO;1$a?`dyh+DP*PsdGbvlY5sMdX!hV$B=mEP0 zaA@}N`=214z&%U4$?Xr;DU;WdY`=_lf*b)U8}8&hW`6ZewWJFibFA|4KI-Gw;-^@) ztU@D#$&oZ|7jgv4x$phP<&f=_ZqcF5yOM2+RGuaqR4S5z?piUxy;cwnd^dI(0_;rj zVd1u27Fku!XO8X~4rKtOQOR|6%!&z6ZIAIYad6CVOlTb(ol0YBA6P4d@>&)$wNt;! zC;A;fjjwYYy~}=nO$HDW6bH>~`O(dHeeXm1MiMD>(W=(6#SgD#%dWW(0%U&$J9eP# z7R-oV-8iRxwWr19McT?qKhF>bL=FAv*3u94$5eerW`jMfuY`|OQzn+4G7b0(3YeW6 zNb81xpxaF{H@3dDt6uMF+x2y=7b_w*j;ufgEqeV8ae>?86t@bWC)AZ)e_GYDIS^bC z)UazAad{72*qf+}?hfU0Oi?xwjo9wj@-K-MHHgCc_si>)EgfT!PhSHgrWI3fTih(M zR-<+?7pM!Vpo+(DEkU8x-0lQZTBu`d{wu}t9MMmJBBY8g+$?7vH-Z;$E|=#L&1%@+ zq0P4THK0sL2~~6Y^9S_gA&y@p>QN=(XnEX7OiH;9cqJr{@|V9&heBQ7F~8Q)D92{E zFT_^c)_EW z38_=e_fgaNcY#jSFv@EG*_Jt%4DEb!KH>K&We@(a5RI_7v84lC7gkLewl)QjYHa(D z#*6+f#0&B22d+19Y-Lf|Q$xa8AwNj%o`+@LaI`(-bj~+>uu@|!cuEQLJJRxxum}o` zMrE9_EUlO$4|`#ky0BT>nCs1~ZC*1KpP3Tv%F@ze8gL>7}jONb}8dk^c4Of{ozvQ$rH^n@S$RZ?Bf<4;(n3R6u6z-0zX6kw zUu2V8$bnikYUh9C4%RjNKNWi_W8w&*tnPQq{a=%yR_~kT_^c=>f>&kJ}a$76>IV?eQ<4o}-FeECE3iyhp!8LAtIsq0=-!?C?5^fR8 z*7N{EV#-#J9DGkgP0Ygeg>*@4YkGR>r-ird4ZqoO`uB`{`hbVI`#w$*=tK4HHfOpM zh^Txh6&w~_8u@ICWU^UyEFp(691H-PMe~Z*_DCVCqLg;S%J0UQs{;h{vw#mr%8nl>qC~U zGosLZzTTew052|^kRF^g0oV3iy4*8~{gPR0MNrhwQ_`isx$hk6yPq-mvC^RtCN%)r zuI3+jBck14SJ=O?X_m(8|C@pv#m78|N(ou55t4iI5Oh7s+FEw^tfoclI1XnXd6W+wDOY>7Tua^?e}CLF<~sqgLjFQrpN=pU=)X~*}l25OWnhNLDE*e$Pj z9H19@4DA875=seLxtp_~uuH>wrd4gh?(|2mlA-}#k^j>DcU4MUG-d~0Q4aeMkSI&Y z)|Yu+xJMyxRVON#5<&G{je&BQ(-UYx^jyog4Rm4ayv%cLdBd+?GSOgKLJ}qJN$7>C zBxYwFqK$1QWG#CUB2&a>0S=W!Ia~jjg7KI(Tk$I@sx)uzT9{Fmj^7I$P*v3I&JkuX zYD7xDbQR)kD=&;ctluAX7o?*o`dk@v3mV?0?&OwTwchrMMVyjRyZZw8ib}A$$d`4_ zrLYu2cOL0zji0i-Rc{e_EK>5Qyjo%wZ&TdVt#bC$at`KP{@A&Z1Q@k_#;qj zZUbLNY~Edql0j+x@KaLh;FOdaTKqDH*$_MB<4N?mWyhvPtsCNkDV<bFGJ( zGfE2`k;)p5WgbrJ))(bV0e%^hY{nl6ytbUNdj`i4$~*6*(xS*g z=B(XtJ0ynb!4J~jkAEI4B%L-C{~9VjnY!wCR7fN>cy$nt`cvyX;?0Kysf)k)5>sJb zoa>?X==K>-&@C;Gj&2xfh91=&AF&JI?6<4h8kS)1kL?4?(m0gu=WFIH^z84x*oJFu zHaGr?jh~Np%Lkz{a;&`H|8wXYQi4Ia8X8jU^rpu0k+{}9FCn|&WCT4td=+7YB^2GWYnj ziQ0AzK^!%TxiAUW^W4AN^>MX{c`w&@wix;Um;;(}I8;Ra<_ui0Yk|)^e#Xb< zSTARSNy1Y$;4P<(q9;UNL7VqpY~tEuOEz!cVn#vse(Yulaf;^u4A}nLX!jw^s21m1 zoebD`bTG1*5?x3wvx+`@>lXuW9$0%GOYqV?ak$05g0?1We9>@xA@>!;K0G*pQR=#UpcTS(tm zAi&tVfM@Q_Rv(*>)Z@mrb}r{n!T zZ6Ho#z$*nQ(4oMIhG6?|K)~ZSRSSwOe-h8!3DXolv-bxBppIr~+?#|+*9s^(`PGMR zR@NibtWy&24`daU(C3mbm=`7%eT3BXl~l_{r5DoV7ZhXQp(u-vxgKhSw2>8ghJ03L zr0U*#n)82OOaNJmGUya`cMlXUJ!8MgreETC_MXLlOWLRo2v-sXgNj0^Vs-@cTT=4k zjU<9!H!-(n-IkKaNpuL1V`b~`Q0^B&@4&6su!z3Xp38T6emhI4a zPWkZcW{pqy#3M?sj{;LRF#z(!ht!{oP*|~fjmR}A7g8H>eBIbHYaAGlwutl=zjxGH;a=528s9mFat zqV>zc%%dfs@ZFMRh?n)v>B}>J&E1OxP0FgIT`~C;P*YOVNvCPEy?McLp%+^xY`u&2@Pe8n?e1d*)7aiJ+ zNl8p_YYQe^|DCiA+){+#>tPGY7NJ=cDsU+!AZZ$QSd4RC5}6PomUnEUI?d-t}lTMiSN zy2XrkMfQGQg;GQOLSt6rFI_aiyq0VEWO4t~Xq?sQ+b#l+c|50CVE5qFsrpFxQ|2AR zOW$rQ8NGH=jo62k{f@C#zX5-JCgGR|*Xu`i%&V`LI6?pArV|!v=xs}ExxhhiS4|cz zqKi!;2o|}l=-}mC&*g9_9E*h2l{tf-YI1gl1`$VLd>y9;5{Hj9S*9QGD;@q|K={p& zyG5ukqX+#nl!`2C7Qo7*<;^cF1q=|*a!@H%DN>0bP#uS!X*ih(`I6PI3nj4$;0v+S zegavR{otIs9MSNl)^#`mKfh^|)oCC?N}BKx9sH|o^-=^b-`$BE8Mt~J-*(J}T!~Wi zc%I1$hy1-}#J%MWa)3=kfuNOemszd(6WoU;H1?W<$~Ue9;^XFSzXDDm(^iLv2H~vJ zkY+0GqF@IH(F`OXEYy(wpbq>s!n delta 6131 zcmZ8lc|6qJ_x}uKOZMes8^hRA_B~`@2NP04%NQe5w#HJ@Ld#@}xVBK1vLxAxq2x^3ObtPZ*1=)^2!inE z{=25AKcC*qA$MOGk*oNdTTL{zN};X;U@0hbBL9NKHRwGbSy zBwmwHy7z&}7Rsg6!oMDOni>eo`;OW?PCLOp{^tDM{l&@kba0VFeHGttw0~f51bD%* zk$=_l|8Y&gjMMm!AL7M8Jg1X}j-jT>T75l#ZGUY8DpfyF-$XYsz#xF8tsQ6>XlSHk zq_1tFZK4wpXdFPHfEmt1>N;y_RDV7FKw|?M)kK%3;NjA%M0 zCN!f!L&GcQRvg5+r=6v~MB)&{CoH$7Sj@xWih=mKlF7*hS2~pu77-R0skxaR5k(2M zrbN)Gnl@n!nr1+7Fe_$5cQoP<0uI?D-+y5thHg{fGc9whIs=7!pEhr8^WANK=~6WQs3&+$SRo;s%o$W^o6?ykD7gOb4p7V^3c9(4QA&$aBlLCe$=O?; z1P806nYR1Xb>ednfd3t(eBSwFm|*cdb!5?7^s59}A$2CdJC#V%4!0Kp1m8@va2-gT&+ z(2XRxxQ~x1Uy!9K!S=l{K9d|R+o9+PC5)KaQ`b-iF1gn%ku zNlE;VZYW=;`m^?=Y(FV(>{z-4=ev!Ve41An9WULx2)&C)v~1pLR&SjnhZ+`5b}C>1 zEj}^yYe;Gv@{*N9u^RE&O{i<~H{bqD&bFVxmrv%1)tj|hkY1Im)M8@joJ&(JU%~NV zQ*Y0hh)7xpD-}Z{!h)Bwkt$924o$__TR8S`O*mC<;ln!?Iv!XCFXwHsnfVup6s84F z+Kw1G7x#$tuF10=VD*w;Tu)87{G>K0_CR1l|Tk^Y=`Bn z4B5Fx?eVjW6sKMmKmlE>{Kka&GLS>M7vS1u-~9dlnFB`um|@y2{K9Bj=jIx?0{l)v zLa~{Z{iT0?eEVCE!HEHW@Z-@kuGb$ReJS@d-ie2Ub}i9f(|j}guq7&O3JVL4V22cp z;)@}0XZ4{MbGgoz`N_BByebYQVN5ZrMbO+h(g|pa^TNj$9}>SfUK{zbc#*1DfB_m; z3z*AQx{pD?_Hc*6U%I2@qL<~#_a)up*>V3J@_7Ea0a$&^{`P0;qgAAHf*;yvI)N>~ z35ZKJ?2z~heGt2?^=eoLAzu-bGhbIW12hB_T+<8oH^NY)9u4|SYU?Y9%Qd_GH$Hjj zu(Yc&p_C9-JM@SOjiiSx<)Aks;0h#E)!zQ~`ARcVZUl8r!{FfyAYVWl6;PWWg8t5} zRIDpJ)nj#&Ebtv2F?|9a2}q#~Bd?E8Q){J6>(Nmgm$L+uqx~E5AA&^zW%T~WNvsaC z=Dw^WC6wA7?V3KG$rJkkmIU8m*KUsmpX})u1cV2-|K^2tgnUw>D z#OXa8yJCVYPx)PMyflCnt#iGhDCdUfui#=Y(TNPJAe=#q zh@dZ5!mMXS5?q9#MYH`IXAi4_)Q6*$59@XS86j1)kSBWv66*7|94;ADv`k$js(zCh zXa_z*N|wMZy#WdtPcAm_aO#qoo0;Smn=O0(v-W$()@FUc7Hh}%BH!<-1NpLvKo-a0 zjwbL+NF4MEiJ>yz-D;r=v>jg@zTQ1x-AvlrZ~oyJ$^yusJBRZwKr2E=4SV$CnT~m5 z0{TYc15%i_gRl^qthGH60-^~oFHe>gyEfS2p3S^Bm;zzKD(G`X;%z9@_}Fw@!Y9q~ zq?s}ywMzdd)K=L-iKGV680`GDx-n7;z3@o!H~pTpU|;sqYpOQrhBTNECZIv|Yww^? z@L2ZTNtN%eS4*AT)pIKT0t697)?t;4m56fnzaPCLb8SIlMc=lnj7i`lqGZ4GnA36S zK}o@nyYt30(&Z<$ryaXLFSR_@o@}It)0T?JX1e(kw8MMk+4M|im3@WIoe8pexhQBC zQL(P)^_hlLxB4eigF`D^&yw@9L~c%PWCz@Z9=eykmV(Pp4KUYB=dXL%_Gyvok+ByO zh6^rWohbfDrgZ8sJf3VkV#Z!?wI`pvljl(z{jcPoG#e=lN^meOn9WBC$6bef*~YvX zq6O_%={Z-wzCXF!{ZAG5C>kS%;=kO!iI04+*2Aknl~^6rNh);@xGAj=745}Va;1fY zF=E&T+Wf{!U{ycz7Yf`{En!vE8fIOPclqd)Y1B0bPpTZdI>~>2}mLGDh6h8PP zjH$IA;Kd}+iuJ@PI6(;Mnp*i+PP!$?Q9o0f`|5# zcez*lgCk=2BbB=*u~L2@(DmS9IKOk=1^t8Cdsm74Vc(TPUkXQhMMXRe$BP?}FK*uM z>2xx0_o0i)Q^J3s-6$b64;t$}BO5Y=;4bAYdx~;J5qu&NCw=z@b z`JM+8tNzgNq*0?7^vI=+{}euU25vmzy1+QuAn%+_wO_UXk-jMzkRQKGNcT&!~+|VdhM;eHikRo~WoJoMHFFG}Jix@xLPHm0dQA#Y;`oq{3n!imjoff*3 zJx>mR+i*pvT8NtayN%r9^);MKZ}Jp?HxkOuxkh%_Y~*vvLfBoHWOJ?P&P{>5D!KpV z+sFvpzvMhkcKt)>-ehLuxQc=nxwJ8DqxM?<8IUF^0d`BSMi<22j6y*D4UhOkCB-hx z0mDrlPxC}r09pIYqt)l2GoJL|(0l3jj%UcCM*=|a@7-W+MpOumL8mT-cKhf=E2J`W z+OgqpPnu;)zsib!ZQs@4lO*1$A0g+%jopG)q}+C9SXVqTFfiK`JIQvP?7y2vjdTqQ zST5)!F`jB@yQob&DWWSrK`M z&E`Q13}(^&m)41bfrHRB<5t0{11XYj7p9s6A5Hky0dpxy_T0&ld%FZC##g*}t9fa^ z|K4=y=EJZ{`75DxNg*qW_0o*AGmN_{y1vAkWgW$T90xES&>mEYjK)mT3 z&Y$~?9kRT9OgsAy&x6k_&COeyUO<~?UX{IGnYG#K;=;H_%!c{bKtoy;?HgWr0%lPq+Ld{5UW7tpu@Hp3y3AY;G zAv^hzol|#6v>TI-WkuIG_jzzFfX6ZfbRcURwr7rqM9SrlTae>wV^{u=^Q8iASvgc) zYWM^+BW<*8b<*n7q^2*6B74AW9JnPb&)!g{6ZWj$fJ0p5!#O(7)wmYz$VxGn>+XdR zkSi;Uej&_tz~J8Q`cTplSLs;a8nRDk(Km$!Fh;E;rLaR$Z&UVk``iGT$?>)^SBR7f zgyjfmjD^Z$gusa$|L5te+hJ+8iS#nAaXmvD;3OxDij6$}2f7wlnoui7SGR8+k}t^Z z9zF~X$>C9VQ-wiT25#?VXBIgJNXaUH@od}2697KS$$@S;F7&Em#VM$*RNSk~;O61<*v_918dVJyyLZOKJSBU`SBKg}=ZNKrDdqhQ4A)8wJ$ZXN z!YaP>X^~H$6enh7g|skw?b-m=<|PE19yEQq;8LB=xSvl{)r0^=?~tWarjGmnA!I&KI1Fz&ddt%++P*}X}d+z<#a5pCJ4PL{kN%un}GD}wAgR3vj^5&ziV?@DsQn{3$ttcQ|G{P%mXGt5dC|?X$%^m1MX1p5LN_rc2SM` zZqk;Ct(2^Ww!s^qR)K&<`u|9Uj)VmG?wAS6C!G`=PUCz26weY$t8Af%288Wfiu-v* zJM2Eclj)UiuAt}8vRhg8@Gk3BKuwY8Rw3T|81kIz$*xk7-%ZNz(_DFFM167DB}O!z zO8ecMS?3k|?CmYLz~=ly+fxtT_HdJ~nwiVH=~QZhGm5fkEhB6aQbrO*c4}q*v_84# z5_9iSohk50Q5l`e(8cW8wC~G#KF;S@Xw+l&bJu}tz(|Q6w?CmGN-M6g{XAA=WcQDs5R)7bjL13&x1b=!xitPMD{c z9(0X${wj7Y+{ST&dBocQ$SV;Xx;Slf;o>HDEPC~{&6Cm>1o6*%WN$AS3S!VG0sqZG zNa&?p$d%B8(CJxDHm7&2>6^L&qZ{`-7yWWMiSD+#=_h^`v?-~e zZZk3CP%|K-3pcXGz@ajUKt=lJl~}^1W9zNe0v5}I_ft9!=l3{f$!8{iU*B5;A;4K# z7+o=UAA9#2(!D{{Xt#Gtf1S4C%EQ`E7$8Ym5H;&B%!D2s=g}D-%^S5p)1@pWnqtTW z>Xh+l0>uyOBsJPWKc5=Ev?)V~vBo0nt5`xwu&BHWmC;bdhP=D7a%7#CYBtQo;Dc6q) zpC?xrOdd0>z97xsl>hS9H8UUUu!|C_B>4(>iO9 zW|~-?p>HPX$o1wdvHVX_MOT5}U9{m9KlwDjYr-u`p(A=Mj}88>k9sz>RzSx&zPhw< zm4~~OT*Jp|XuMyF0|frQ8hK)moNn*;U}xytI+M1JV5&_6SN{I#BssA^cs*DdMq_MY zjj7AG!|u!z!z>essyT&>7b~_X_5MFxti1Su& zgQ)_f!{6lB=-FOXHf0b-Q{IB`H3XajNMF-vaR65yk1H?0m6zel>v83`apk?Z@_z+! p{5Tt2Tr)1-6&D}QA;ZObmX9C92_Y<+n*RhVJ>F~BduZe)C6A&VWg-`cW#zsa( zsx{Kw)bXC7nxa^ZF}g?{7gHD&wXP+kF61N7>56h$xu7DY&~|S*toHcM7##Viq?6&Oy|ilrrr4{y4#8cTb^b#yCM%UL zUinpyd-mhK^~je}uQe-i-2T^%MS@E2PeY(o*&nIIfOO){APc>^)aCQ|ro&PB1Qt(G z$jMa>hlG17@pvmwMEivI>r%(vhf4I@`Tay3wm-bT82jw{2Vqzzl{~o2m@-$&9JT?! z*}oZyM3eM*@X;_>87|Vc2m~RFH%6dEA(R%X zp+hTwQ;l@Mk1j+b4Mp^v23l;PR1JcW!V5K!BFeM<$4O#)~#k$OjUz7I>HDB?*Z#+MI+HLa8eWgqu{5 zj8X*BrDT}U&XZCQ>xPBgoPk6jm^aVCA5Mw)^cVsT^SuT@E~=@|#yaev(QMp6Bk#?D zg5hU=eLf6wRP&IcXy6n18)|&Tj+TuDdw5G3CbMx{zj7vxn<*l*dEYJ^299#8 z6?>s6>v-z?)SC2SG&Pobk+J%mi+N{^X?@02IGMWSE=sQyBQemmO00)BhH_4wG`lNsm@=1(s_7! z3nU8MnC5U~+D3N1lvzETQh!=H2j ztVFDmJ7XzGAw6SSrZUhD#!O<*?l6{zOiGorbjGr%QqC6PGrA&Y3s6A83YN<}^|VdF zvd~WL3bq)<Z;Di0PVa_T%-ER_;NjKbN>$(?pq!J delta 1540 zcma)6YfzMB6yE3EdvW`(KM3S-V%jW#bb!k7~kZd9+;Q6Z>6DX3$N5~bFtuG#qDU||=-0c%nBc=%XryZ1pMy?*Y12YXs5 zUAXqE5_hb}dmE81#r3r~OHT>Tto zcTvdKD`i>2A8z9FMyBxg3+;TV^Y@3&=(F98SY_tNuYS)<}g9t}-U4NlA|^ zz)kDL$6+Xuo;}_UT#+nKHbJ0APcHzmQrq(;<)oM-d$&t^Y#vjfFj)GF^D3eh#o0nk zvKGl&oRAL5ohtmsO)9jELIX|m#A`6|I8Rs@?ogK({LxM}FHFT|?&^(Xh74LVPQbso zc^s~fK>;NPVyo--qmrSSN<-lXf4(;qr&I`` zx7CQi3I4Gf34jNk4@W$*=oJmL*g$a_1VG1)8c0gq<`Z2g{gh3SZ=@+c&sk#|LjHn5u8<{%enqLdhsqJ#=!uuNp?jsc;c zvSLw$v2-~WCbaUHxzIUb0WY41XkZF&NyMM_6#MlU0?fR@04PN@jSI03W_q#^1K7=b zlc8d`#&0fyL5V6Jn1(vwAW7-az=!Uo<137(T^XXd2b^bM&>e~V^ED9Ll|buqa2G3R zQ7$C-ab^@XLc#s>z?o3!dl&!U82!8+G5CqRP1uPEbjXBECf^|0gtM-v`d`>ro70c^ zn9F>BA&(6>BmRB&1_ZjGhlbum2d>b`Vx+*!O%~0nhcC=)~*~3qjVGuRwh3mu>*&lSurAMvJ1;(0|{hy zVRzwAD_z+v3{$-;OUGw)%TdD*V>2<#>)lw1J4(6Ahoyu>tHU*6+Hh^S n&M$u6rhshM8>AyU Hu_#R#+jbL# delta 49 zcmV-10M7sQF!eAfP)kQa3;+NC!Bpo5k@Ew!fH^@wmnG%^N+10K+#EvtxdVU%snDq9WMcQnzhh z!V|+}_=3xH7Xzbj029#O3?QxmdZ&UP$bbSyAT12U>OlMeDEDBpA-}THLYB=>N~6#E zHcn!^P;{hiI!Kj%adJ^&NpgmMQEEYc@#F%2Pa~)@-h)tOAj@O}v|hjO*Q)nC|FlgX q7^pDd31TZuR1^b-9+QC9W<7x|jGOs{ez7ve*liY+c+WV|fE@rTm1Q^p delta 353 zcmZoLX)u}4!IBuh@$keR8AiQ{E5rG2>M}6+GXQ}D1A`6gL`7kq1O^5kApegJ14GDU zMJAoek&Nn$CX*W(>lt|`GcpPDL~t`OFalMu0ny|{0fWf_Omi6>C%xT`3v#S}adJ^&Npi;IdHkM6PzBYTPz4}E z|9?7mXzuiLYpgyR^)u=LbwGhMh^;VDQ4HvN0|6~|28REt3<_+=9$L5-QWE4e&#*j^F8M|=RD6j@3Z8R>Y5!UZ>^3XN`#4 z*)PA3^6Voe$o-`81A(=WdnJBuO&yHE$8oEwBX+{bD`g17cH66!yclc@s=E!x-F9(oX4)?<7L6A z!L);55bMOx_wS3u4xrDr>GwBDI1s~TvwowYw!Ws1p|7!?uC9-+kAc3Pub;1uzs@GI zKgCy1*GON_*hoj$2maUf1v6}M>-07C{f+d=dj7hbHd1sp`5NmOl6}Yq2D%iop`n4H zp0S}XS#KjlQ|IzY_E-%1I2!Fk5)tHu=ke%6u}QZ1gX~R8VJ}PpnNu8W<>W;>*Qn$o z@)+d77TsF^_c*h{fGi}}v|FA5kU@ojdVVk*o%V8Ym`w;bp3a@79OSQ zBwaoa`(i(LxqYTJ-Qe>YNfepaItr_;4{OhUm-i{YvYpS1@4LkckmpiDKekvuLCi2n zjBUdjckdayT#uRECg*wMfhU(TN}H7|g0v?nTRAYMGLGjkr6bMCGxDH-O9hQ*e?JI? z$EG|tDdOpbiWrmBT>XljV1Qd3OmYdM8+?}gkPZat98p_x&Fmwccjjuz6Os#O0@iVh zqk_wO-@qk3)HeA0a@4SS8}C5NO{H2cCLn~H6QxvLD1`u0=e=rF=_cvA`^U*w`IT`@ zz%BHAQdyFAhL({(}FG2s3n)`6_#ILIh-V%i=3Ff+6T7tNGFu}Ck#T<3d8C*FS|vX=VrL1 zHxGTk!O9IIleEwv3M)5Vnad}jscltU5iWRdV!cuD#PvbDC;xyWR*@m>axD-kfsec^ zBSo^Q)Y$CL2MGIENDFc#S=&4`I2cUvaHBdlDoscU773`9@|b;+K%(z3eX2+Fcn|n^ zl?g4i?$4lutq)ZRw|$1}3Qmizd|!+;_?_uj1^Waxet#M`$SL{>X>@Myoc-chWuKgceW8gB`~%eS%Aw=E-E~lSIHWT&JuA+vu*)De$pvtF#(&F}?g_>70^?c^;7}=B@i5+U{ zTucK2dDO0}fZF3`S(IQ9Y+emT^9Pwp9L^R=5Azsc?}QcCJp zWm9UvHn5vt1wBO5X72XQPk35CZ2MuJep+g~6qi;SIL)t+ARM)60@@UgD+oT5$mU2F zDXe)dxD3fY|sx6${WcbJ)?K#NN_s_@4F5xLcXN04KvApxw45J!85pfCK} zIl@QwbXXMknHszqPM7>Yuu}vzm>LO;1$a?`dyh+DP*PsdGbvlY5sMdX!hV$B=mEP0 zaA@}N`=214z&%U4$?Xr;DU;WdY`=_lf*b)U8}8&hW`6ZewWJFibFA|4KI-Gw;-^@) ztU@D#$&oZ|7jgv4x$phP<&f=_ZqcF5yOM2+RGuaqR4S5z?piUxy;cwnd^dI(0_;rj zVd1u27Fku!XO8X~4rKtOQOR|6%!&z6ZIAIYad6CVOlTb(ol0YBA6P4d@>&)$wNt;! zC;A;fjjwYYy~}=nO$HDW6bH>~`O(dHeeXm1MiMD>(W=(6#SgD#%dWW(0%U&$J9eP# z7R-oV-8iRxwWr19McT?qKhF>bL=FAv*3u94$5eerW`jMfuY`|OQzn+4G7b0(3YeW6 zNb81xpxaF{H@3dDt6uMF+x2y=7b_w*j;ufgEqeV8ae>?86t@bWC)AZ)e_GYDIS^bC z)UazAad{72*qf+}?hfU0Oi?xwjo9wj@-K-MHHgCc_si>)EgfT!PhSHgrWI3fTih(M zR-<+?7pM!Vpo+(DEkU8x-0lQZTBu`d{wu}t9MMmJBBY8g+$?7vH-Z;$E|=#L&1%@+ zq0P4THK0sL2~~6Y^9S_gA&y@p>QN=(XnEX7OiH;9cqJr{@|V9&heBQ7F~8Q)D92{E zFT_^c)_EW z38_=e_fgaNcY#jSFv@EG*_Jt%4DEb!KH>K&We@(a5RI_7v84lC7gkLewl)QjYHa(D z#*6+f#0&B22d+19Y-Lf|Q$xa8AwNj%o`+@LaI`(-bj~+>uu@|!cuEQLJJRxxum}o` zMrE9_EUlO$4|`#ky0BT>nCs1~ZC*1KpP3Tv%F@ze8gL>7}jONb}8dk^c4Of{ozvQ$rH^n@S$RZ?Bf<4;(n3R6u6z-0zX6kw zUu2V8$bnikYUh9C4%RjNKNWi_W8w&*tnPQq{a=%yR_~kT_^c=>f>&kJ}a$76>IV?eQ<4o}-FeECE3iyhp!8LAtIsq0=-!?C?5^fR8 z*7N{EV#-#J9DGkgP0Ygeg>*@4YkGR>r-ird4ZqoO`uB`{`hbVI`#w$*=tK4HHfOpM zh^Txh6&w~_8u@ICWU^UyEFp(691H-PMe~Z*_DCVCqLg;S%J0UQs{;h{vw#mr%8nl>qC~U zGosLZzTTew052|^kRF^g0oV3iy4*8~{gPR0MNrhwQ_`isx$hk6yPq-mvC^RtCN%)r zuI3+jBck14SJ=O?X_m(8|C@pv#m78|N(ou55t4iI5Oh7s+FEw^tfoclI1XnXd6W+wDOY>7Tua^?e}CLF<~sqgLjFQrpN=pU=)X~*}l25OWnhNLDE*e$Pj z9H19@4DA875=seLxtp_~uuH>wrd4gh?(|2mlA-}#k^j>DcU4MUG-d~0Q4aeMkSI&Y z)|Yu+xJMyxRVON#5<&G{je&BQ(-UYx^jyog4Rm4ayv%cLdBd+?GSOgKLJ}qJN$7>C zBxYwFqK$1QWG#CUB2&a>0S=W!Ia~jjg7KI(Tk$I@sx)uzT9{Fmj^7I$P*v3I&JkuX zYD7xDbQR)kD=&;ctluAX7o?*o`dk@v3mV?0?&OwTwchrMMVyjRyZZw8ib}A$$d`4_ zrLYu2cOL0zji0i-Rc{e_EK>5Qyjo%wZ&TdVt#bC$at`KP{@A&Z1Q@k_#;qj zZUbLNY~Edql0j+x@KaLh;FOdaTKqDH*$_MB<4N?mWyhvPtsCNkDV<bFGJ( zGfE2`k;)p5WgbrJ))(bV0e%^hY{nl6ytbUNdj`i4$~*6*(xS*g z=B(XtJ0ynb!4J~jkAEI4B%L-C{~9VjnY!wCR7fN>cy$nt`cvyX;?0Kysf)k)5>sJb zoa>?X==K>-&@C;Gj&2xfh91=&AF&JI?6<4h8kS)1kL?4?(m0gu=WFIH^z84x*oJFu zHaGr?jh~Np%Lkz{a;&`H|8wXYQi4Ia8X8jU^rpu0k+{}9FCn|&WCT4td=+7YB^2GWYnj ziQ0AzK^!%TxiAUW^W4AN^>MX{c`w&@wix;Um;;(}I8;Ra<_ui0Yk|)^e#Xb< zSTARSNy1Y$;4P<(q9;UNL7VqpY~tEuOEz!cVn#vse(Yulaf;^u4A}nLX!jw^s21m1 zoebD`bTG1*5?x3wvx+`@>lXuW9$0%GOYqV?ak$05g0?1We9>@xA@>!;K0G*pQR=#UpcTS(tm zAi&tVfM@Q_Rv(*>)Z@mrb}r{n!T zZ6Ho#z$*nQ(4oMIhG6?|K)~ZSRSSwOe-h8!3DXolv-bxBppIr~+?#|+*9s^(`PGMR zR@NibtWy&24`daU(C3mbm=`7%eT3BXl~l_{r5DoV7ZhXQp(u-vxgKhSw2>8ghJ03L zr0U*#n)82OOaNJmGUya`cMlXUJ!8MgreETC_MXLlOWLRo2v-sXgNj0^Vs-@cTT=4k zjU<9!H!-(n-IkKaNpuL1V`b~`Q0^B&@4&6su!z3Xp38T6emhI4a zPWkZcW{pqy#3M?sj{;LRF#z(!ht!{oP*|~fjmR}A7g8H>eBIbHYaAGlwutl=zjxGH;a=528s9mFat zqV>zc%%dfs@ZFMRh?n)v>B}>J&E1OxP0FgIT`~C;P*YOVNvCPEy?McLp%+^xY`u&2@Pe8n?e1d*)7aiJ+ zNl8p_YYQe^|DCiA+){+#>tPGY7NJ=cDsU+!AZZ$QSd4RC5}6PomUnEUI?d-t}lTMiSN zy2XrkMfQGQg;GQOLSt6rFI_aiyq0VEWO4t~Xq?sQ+b#l+c|50CVE5qFsrpFxQ|2AR zOW$rQ8NGH=jo62k{f@C#zX5-JCgGR|*Xu`i%&V`LI6?pArV|!v=xs}ExxhhiS4|cz zqKi!;2o|}l=-}mC&*g9_9E*h2l{tf-YI1gl1`$VLd>y9;5{Hj9S*9QGD;@q|K={p& zyG5ukqX+#nl!`2C7Qo7*<;^cF1q=|*a!@H%DN>0bP#uS!X*ih(`I6PI3nj4$;0v+S zegavR{otIs9MSNl)^#`mKfh^|)oCC?N}BKx9sH|o^-=^b-`$BE8Mt~J-*(J}T!~Wi zc%I1$hy1-}#J%MWa)3=kfuNOemszd(6WoU;H1?W<$~Ue9;^XFSzXDDm(^iLv2H~vJ zkY+0GqF@IH(F`OXEYy(wpbq>s!n delta 6131 zcmZ8lc|6qJ_x}uKOZMes8^hRA_B~`@2NP04%NQe5w#HJ@Ld#@}xVBK1vLxAxq2x^3ObtPZ*1=)^2!inE z{=25AKcC*qA$MOGk*oNdTTL{zN};X;U@0hbBL9NKHRwGbSy zBwmwHy7z&}7Rsg6!oMDOni>eo`;OW?PCLOp{^tDM{l&@kba0VFeHGttw0~f51bD%* zk$=_l|8Y&gjMMm!AL7M8Jg1X}j-jT>T75l#ZGUY8DpfyF-$XYsz#xF8tsQ6>XlSHk zq_1tFZK4wpXdFPHfEmt1>N;y_RDV7FKw|?M)kK%3;NjA%M0 zCN!f!L&GcQRvg5+r=6v~MB)&{CoH$7Sj@xWih=mKlF7*hS2~pu77-R0skxaR5k(2M zrbN)Gnl@n!nr1+7Fe_$5cQoP<0uI?D-+y5thHg{fGc9whIs=7!pEhr8^WANK=~6WQs3&+$SRo;s%o$W^o6?ykD7gOb4p7V^3c9(4QA&$aBlLCe$=O?; z1P806nYR1Xb>ednfd3t(eBSwFm|*cdb!5?7^s59}A$2CdJC#V%4!0Kp1m8@va2-gT&+ z(2XRxxQ~x1Uy!9K!S=l{K9d|R+o9+PC5)KaQ`b-iF1gn%ku zNlE;VZYW=;`m^?=Y(FV(>{z-4=ev!Ve41An9WULx2)&C)v~1pLR&SjnhZ+`5b}C>1 zEj}^yYe;Gv@{*N9u^RE&O{i<~H{bqD&bFVxmrv%1)tj|hkY1Im)M8@joJ&(JU%~NV zQ*Y0hh)7xpD-}Z{!h)Bwkt$924o$__TR8S`O*mC<;ln!?Iv!XCFXwHsnfVup6s84F z+Kw1G7x#$tuF10=VD*w;Tu)87{G>K0_CR1l|Tk^Y=`Bn z4B5Fx?eVjW6sKMmKmlE>{Kka&GLS>M7vS1u-~9dlnFB`um|@y2{K9Bj=jIx?0{l)v zLa~{Z{iT0?eEVCE!HEHW@Z-@kuGb$ReJS@d-ie2Ub}i9f(|j}guq7&O3JVL4V22cp z;)@}0XZ4{MbGgoz`N_BByebYQVN5ZrMbO+h(g|pa^TNj$9}>SfUK{zbc#*1DfB_m; z3z*AQx{pD?_Hc*6U%I2@qL<~#_a)up*>V3J@_7Ea0a$&^{`P0;qgAAHf*;yvI)N>~ z35ZKJ?2z~heGt2?^=eoLAzu-bGhbIW12hB_T+<8oH^NY)9u4|SYU?Y9%Qd_GH$Hjj zu(Yc&p_C9-JM@SOjiiSx<)Aks;0h#E)!zQ~`ARcVZUl8r!{FfyAYVWl6;PWWg8t5} zRIDpJ)nj#&Ebtv2F?|9a2}q#~Bd?E8Q){J6>(Nmgm$L+uqx~E5AA&^zW%T~WNvsaC z=Dw^WC6wA7?V3KG$rJkkmIU8m*KUsmpX})u1cV2-|K^2tgnUw>D z#OXa8yJCVYPx)PMyflCnt#iGhDCdUfui#=Y(TNPJAe=#q zh@dZ5!mMXS5?q9#MYH`IXAi4_)Q6*$59@XS86j1)kSBWv66*7|94;ADv`k$js(zCh zXa_z*N|wMZy#WdtPcAm_aO#qoo0;Smn=O0(v-W$()@FUc7Hh}%BH!<-1NpLvKo-a0 zjwbL+NF4MEiJ>yz-D;r=v>jg@zTQ1x-AvlrZ~oyJ$^yusJBRZwKr2E=4SV$CnT~m5 z0{TYc15%i_gRl^qthGH60-^~oFHe>gyEfS2p3S^Bm;zzKD(G`X;%z9@_}Fw@!Y9q~ zq?s}ywMzdd)K=L-iKGV680`GDx-n7;z3@o!H~pTpU|;sqYpOQrhBTNECZIv|Yww^? z@L2ZTNtN%eS4*AT)pIKT0t697)?t;4m56fnzaPCLb8SIlMc=lnj7i`lqGZ4GnA36S zK}o@nyYt30(&Z<$ryaXLFSR_@o@}It)0T?JX1e(kw8MMk+4M|im3@WIoe8pexhQBC zQL(P)^_hlLxB4eigF`D^&yw@9L~c%PWCz@Z9=eykmV(Pp4KUYB=dXL%_Gyvok+ByO zh6^rWohbfDrgZ8sJf3VkV#Z!?wI`pvljl(z{jcPoG#e=lN^meOn9WBC$6bef*~YvX zq6O_%={Z-wzCXF!{ZAG5C>kS%;=kO!iI04+*2Aknl~^6rNh);@xGAj=745}Va;1fY zF=E&T+Wf{!U{ycz7Yf`{En!vE8fIOPclqd)Y1B0bPpTZdI>~>2}mLGDh6h8PP zjH$IA;Kd}+iuJ@PI6(;Mnp*i+PP!$?Q9o0f`|5# zcez*lgCk=2BbB=*u~L2@(DmS9IKOk=1^t8Cdsm74Vc(TPUkXQhMMXRe$BP?}FK*uM z>2xx0_o0i)Q^J3s-6$b64;t$}BO5Y=;4bAYdx~;J5qu&NCw=z@b z`JM+8tNzgNq*0?7^vI=+{}euU25vmzy1+QuAn%+_wO_UXk-jMzkRQKGNcT&!~+|VdhM;eHikRo~WoJoMHFFG}Jix@xLPHm0dQA#Y;`oq{3n!imjoff*3 zJx>mR+i*pvT8NtayN%r9^);MKZ}Jp?HxkOuxkh%_Y~*vvLfBoHWOJ?P&P{>5D!KpV z+sFvpzvMhkcKt)>-ehLuxQc=nxwJ8DqxM?<8IUF^0d`BSMi<22j6y*D4UhOkCB-hx z0mDrlPxC}r09pIYqt)l2GoJL|(0l3jj%UcCM*=|a@7-W+MpOumL8mT-cKhf=E2J`W z+OgqpPnu;)zsib!ZQs@4lO*1$A0g+%jopG)q}+C9SXVqTFfiK`JIQvP?7y2vjdTqQ zST5)!F`jB@yQob&DWWSrK`M z&E`Q13}(^&m)41bfrHRB<5t0{11XYj7p9s6A5Hky0dpxy_T0&ld%FZC##g*}t9fa^ z|K4=y=EJZ{`75DxNg*qW_0o*AGmN_{y1vAkWgW$T90xES&>mEYjK)mT3 z&Y$~?9kRT9OgsAy&x6k_&COeyUO<~?UX{IGnYG#K;=;H_%!c{bKtoy;?HgWr0%lPq+Ld{5UW7tpu@Hp3y3AY;G zAv^hzol|#6v>TI-WkuIG_jzzFfX6ZfbRcURwr7rqM9SrlTae>wV^{u=^Q8iASvgc) zYWM^+BW<*8b<*n7q^2*6B74AW9JnPb&)!g{6ZWj$fJ0p5!#O(7)wmYz$VxGn>+XdR zkSi;Uej&_tz~J8Q`cTplSLs;a8nRDk(Km$!Fh;E;rLaR$Z&UVk``iGT$?>)^SBR7f zgyjfmjD^Z$gusa$|L5te+hJ+8iS#nAaXmvD;3OxDij6$}2f7wlnoui7SGR8+k}t^Z z9zF~X$>C9VQ-wiT25#?VXBIgJNXaUH@od}2697KS$$@S;F7&Em#VM$*RNSk~;O61<*v_918dVJyyLZOKJSBU`SBKg}=ZNKrDdqhQ4A)8wJ$ZXN z!YaP>X^~H$6enh7g|skw?b-m=<|PE19yEQq;8LB=xSvl{)r0^=?~tWarjGmnA!I&KI1Fz&ddt%++P*}X}d+z<#a5pCJ4PL{kN%un}GD}wAgR3vj^5&ziV?@DsQn{3$ttcQ|G{P%mXGt5dC|?X$%^m1MX1p5LN_rc2SM` zZqk;Ct(2^Ww!s^qR)K&<`u|9Uj)VmG?wAS6C!G`=PUCz26weY$t8Af%288Wfiu-v* zJM2Eclj)UiuAt}8vRhg8@Gk3BKuwY8Rw3T|81kIz$*xk7-%ZNz(_DFFM167DB}O!z zO8ecMS?3k|?CmYLz~=ly+fxtT_HdJ~nwiVH=~QZhGm5fkEhB6aQbrO*c4}q*v_84# z5_9iSohk50Q5l`e(8cW8wC~G#KF;S@Xw+l&bJu}tz(|Q6w?CmGN-M6g{XAA=WcQDs5R)7bjL13&x1b=!xitPMD{c z9(0X${wj7Y+{ST&dBocQ$SV;Xx;Slf;o>HDEPC~{&6Cm>1o6*%WN$AS3S!VG0sqZG zNa&?p$d%B8(CJxDHm7&2>6^L&qZ{`-7yWWMiSD+#=_h^`v?-~e zZZk3CP%|K-3pcXGz@ajUKt=lJl~}^1W9zNe0v5}I_ft9!=l3{f$!8{iU*B5;A;4K# z7+o=UAA9#2(!D{{Xt#Gtf1S4C%EQ`E7$8Ym5H;&B%!D2s=g}D-%^S5p)1@pWnqtTW z>Xh+l0>uyOBsJPWKc5=Ev?)V~vBo0nt5`xwu&BHWmC;bdhP=D7a%7#CYBtQo;Dc6q) zpC?xrOdd0>z97xsl>hS9H8UUUu!|C_B>4(>iO9 zW|~-?p>HPX$o1wdvHVX_MOT5}U9{m9KlwDjYr-u`p(A=Mj}88>k9sz>RzSx&zPhw< zm4~~OT*Jp|XuMyF0|frQ8hK)moNn*;U}xytI+M1JV5&_6SN{I#BssA^cs*DdMq_MY zjj7AG!|u!z!z>essyT&>7b~_X_5MFxti1Su& zgQ)_f!{6lB=-FOXHf0b-Q{IB`H3XajNMF-vaR65yk1H?0m6zel>v83`apk?Z@_z+! p{5Tt2Tr)1-6&D}QA;ZObmX9C92_Y 25) (kafka.controller.KafkaController) -[2023-11-03 19:38:21,195] DEBUG [Controller id=0] Register BrokerModifications handler for Set(0) (kafka.controller.KafkaController) -[2023-11-03 19:38:21,202] DEBUG [Channel manager on controller 0]: Controller 0 trying to connect to broker 0 (kafka.controller.ControllerChannelManager) -[2023-11-03 19:38:21,206] INFO [RequestSendThread controllerId=0] Starting (kafka.controller.RequestSendThread) -[2023-11-03 19:38:21,208] INFO [Controller id=0] Currently active brokers in the cluster: Set(0) (kafka.controller.KafkaController) -[2023-11-03 19:38:21,209] INFO [Controller id=0] Currently shutting brokers in the cluster: HashSet() (kafka.controller.KafkaController) -[2023-11-03 19:38:21,209] INFO [Controller id=0] Current list of topics in the cluster: HashSet() (kafka.controller.KafkaController) -[2023-11-03 19:38:21,209] INFO [Controller id=0] Fetching topic deletions in progress (kafka.controller.KafkaController) -[2023-11-03 19:38:21,212] INFO [Controller id=0] List of topics to be deleted: (kafka.controller.KafkaController) -[2023-11-03 19:38:21,212] INFO [Controller id=0] List of topics ineligible for deletion: (kafka.controller.KafkaController) -[2023-11-03 19:38:21,212] INFO [Controller id=0] Initializing topic deletion manager (kafka.controller.KafkaController) -[2023-11-03 19:38:21,213] INFO [Topic Deletion Manager 0] Initializing manager with initial deletions: Set(), initial ineligible deletions: HashSet() (kafka.controller.TopicDeletionManager) -[2023-11-03 19:38:21,214] INFO [Controller id=0] Sending update metadata request (kafka.controller.KafkaController) -[2023-11-03 19:38:21,224] INFO [ReplicaStateMachine controllerId=0] Initializing replica state (kafka.controller.ZkReplicaStateMachine) -[2023-11-03 19:38:21,224] INFO [ReplicaStateMachine controllerId=0] Triggering online replica state changes (kafka.controller.ZkReplicaStateMachine) -[2023-11-03 19:38:21,230] INFO [ReplicaStateMachine controllerId=0] Triggering offline replica state changes (kafka.controller.ZkReplicaStateMachine) -[2023-11-03 19:38:21,230] DEBUG [ReplicaStateMachine controllerId=0] Started replica state machine with initial state -> HashMap() (kafka.controller.ZkReplicaStateMachine) -[2023-11-03 19:38:21,231] INFO [PartitionStateMachine controllerId=0] Initializing partition state (kafka.controller.ZkPartitionStateMachine) -[2023-11-03 19:38:21,231] INFO [PartitionStateMachine controllerId=0] Triggering online partition state changes (kafka.controller.ZkPartitionStateMachine) -[2023-11-03 19:38:21,233] DEBUG [PartitionStateMachine controllerId=0] Started partition state machine with initial state -> HashMap() (kafka.controller.ZkPartitionStateMachine) -[2023-11-03 19:38:21,233] INFO [Controller id=0] Ready to serve as the new controller with epoch 1 (kafka.controller.KafkaController) -[2023-11-03 19:38:21,245] INFO [Controller id=0] Partitions undergoing preferred replica election: (kafka.controller.KafkaController) -[2023-11-03 19:38:21,245] INFO [Controller id=0] Partitions that completed preferred replica election: (kafka.controller.KafkaController) -[2023-11-03 19:38:21,246] INFO [Controller id=0] Skipping preferred replica election for partitions due to topic deletion: (kafka.controller.KafkaController) -[2023-11-03 19:38:21,246] WARN [RequestSendThread controllerId=0] Controller 0's connection to broker ThinkPadP53:9092 (id: 0 rack: null) was unsuccessful (kafka.controller.RequestSendThread) -java.io.IOException: Connection to ThinkPadP53:9092 (id: 0 rack: null) failed. +[2023-11-06 13:26:31,961] INFO [ControllerEventThread controllerId=0] Starting (kafka.controller.ControllerEventManager$ControllerEventThread) +[2023-11-06 13:26:31,976] INFO [Controller id=0] 0 successfully elected as the controller. Epoch incremented to 1 and epoch zk version is now 1 (kafka.controller.KafkaController) +[2023-11-06 13:26:31,979] INFO [Controller id=0] Creating FeatureZNode at path: /feature with contents: FeatureZNode(2,Enabled,Map()) (kafka.controller.KafkaController) +[2023-11-06 13:26:31,997] INFO [Controller id=0] Registering handlers (kafka.controller.KafkaController) +[2023-11-06 13:26:32,000] INFO [Controller id=0] Deleting log dir event notifications (kafka.controller.KafkaController) +[2023-11-06 13:26:32,002] INFO [Controller id=0] Deleting isr change notifications (kafka.controller.KafkaController) +[2023-11-06 13:26:32,004] INFO [Controller id=0] Initializing controller context (kafka.controller.KafkaController) +[2023-11-06 13:26:32,013] INFO [Controller id=0] Initialized broker epochs cache: HashMap(0 -> 25) (kafka.controller.KafkaController) +[2023-11-06 13:26:32,016] DEBUG [Controller id=0] Register BrokerModifications handler for Set(0) (kafka.controller.KafkaController) +[2023-11-06 13:26:32,020] DEBUG [Channel manager on controller 0]: Controller 0 trying to connect to broker 0 (kafka.controller.ControllerChannelManager) +[2023-11-06 13:26:32,023] INFO [RequestSendThread controllerId=0] Starting (kafka.controller.RequestSendThread) +[2023-11-06 13:26:32,024] INFO [Controller id=0] Currently active brokers in the cluster: Set(0) (kafka.controller.KafkaController) +[2023-11-06 13:26:32,024] INFO [Controller id=0] Currently shutting brokers in the cluster: HashSet() (kafka.controller.KafkaController) +[2023-11-06 13:26:32,024] INFO [Controller id=0] Current list of topics in the cluster: HashSet() (kafka.controller.KafkaController) +[2023-11-06 13:26:32,024] INFO [Controller id=0] Fetching topic deletions in progress (kafka.controller.KafkaController) +[2023-11-06 13:26:32,027] INFO [Controller id=0] List of topics to be deleted: (kafka.controller.KafkaController) +[2023-11-06 13:26:32,027] INFO [Controller id=0] List of topics ineligible for deletion: (kafka.controller.KafkaController) +[2023-11-06 13:26:32,028] INFO [Controller id=0] Initializing topic deletion manager (kafka.controller.KafkaController) +[2023-11-06 13:26:32,028] INFO [Topic Deletion Manager 0] Initializing manager with initial deletions: Set(), initial ineligible deletions: HashSet() (kafka.controller.TopicDeletionManager) +[2023-11-06 13:26:32,029] INFO [Controller id=0] Sending update metadata request (kafka.controller.KafkaController) +[2023-11-06 13:26:32,037] INFO [ReplicaStateMachine controllerId=0] Initializing replica state (kafka.controller.ZkReplicaStateMachine) +[2023-11-06 13:26:32,037] INFO [ReplicaStateMachine controllerId=0] Triggering online replica state changes (kafka.controller.ZkReplicaStateMachine) +[2023-11-06 13:26:32,046] INFO [ReplicaStateMachine controllerId=0] Triggering offline replica state changes (kafka.controller.ZkReplicaStateMachine) +[2023-11-06 13:26:32,049] DEBUG [ReplicaStateMachine controllerId=0] Started replica state machine with initial state -> HashMap() (kafka.controller.ZkReplicaStateMachine) +[2023-11-06 13:26:32,049] INFO [PartitionStateMachine controllerId=0] Initializing partition state (kafka.controller.ZkPartitionStateMachine) +[2023-11-06 13:26:32,049] INFO [PartitionStateMachine controllerId=0] Triggering online partition state changes (kafka.controller.ZkPartitionStateMachine) +[2023-11-06 13:26:32,051] DEBUG [PartitionStateMachine controllerId=0] Started partition state machine with initial state -> HashMap() (kafka.controller.ZkPartitionStateMachine) +[2023-11-06 13:26:32,051] INFO [Controller id=0] Ready to serve as the new controller with epoch 1 (kafka.controller.KafkaController) +[2023-11-06 13:26:32,054] WARN [RequestSendThread controllerId=0] Controller 0's connection to broker localhost:9092 (id: 0 rack: null) was unsuccessful (kafka.controller.RequestSendThread) +java.io.IOException: Connection to localhost:9092 (id: 0 rack: null) failed. at org.apache.kafka.clients.NetworkClientUtils.awaitReady(NetworkClientUtils.java:70) at kafka.controller.RequestSendThread.brokerReady(ControllerChannelManager.scala:298) at kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:251) at org.apache.kafka.server.util.ShutdownableThread.run(ShutdownableThread.java:130) -[2023-11-03 19:38:21,246] INFO [Controller id=0] Resuming preferred replica election for partitions: (kafka.controller.KafkaController) -[2023-11-03 19:38:21,248] INFO [Controller id=0] Starting replica leader election (PREFERRED) for partitions triggered by ZkTriggered (kafka.controller.KafkaController) -[2023-11-03 19:38:21,256] INFO [Controller id=0] Starting the controller scheduler (kafka.controller.KafkaController) -[2023-11-03 19:38:21,349] INFO [RequestSendThread controllerId=0] Controller 0 connected to ThinkPadP53:9092 (id: 0 rack: null) for sending state change requests (kafka.controller.RequestSendThread) -[2023-11-03 19:38:26,257] INFO [Controller id=0] Processing automatic preferred replica leader election (kafka.controller.KafkaController) -[2023-11-03 19:38:26,258] TRACE [Controller id=0] Checking need to trigger auto leader balancing (kafka.controller.KafkaController) -[2023-11-03 19:38:41,910] INFO [Controller id=0] New topics: [Set(__consumer_offsets)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(__consumer_offsets,Some(vfIIZeo7TYKNeTmKGzMlyg),HashMap(__consumer_offsets-22 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-30 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-25 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-35 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-37 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-38 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-13 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-8 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-21 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-4 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-27 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-7 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-9 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-46 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-41 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-33 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-23 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-49 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-47 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-16 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-28 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-31 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-36 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-42 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-3 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-18 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-15 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-24 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-17 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-48 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-19 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-11 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-2 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-43 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-6 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-14 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-20 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-44 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-39 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-12 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-45 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-1 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-5 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-26 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-29 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-34 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-10 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-32 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-40 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController) -[2023-11-03 19:38:41,910] INFO [Controller id=0] New partition creation callback for __consumer_offsets-22,__consumer_offsets-30,__consumer_offsets-25,__consumer_offsets-35,__consumer_offsets-37,__consumer_offsets-38,__consumer_offsets-13,__consumer_offsets-8,__consumer_offsets-21,__consumer_offsets-4,__consumer_offsets-27,__consumer_offsets-7,__consumer_offsets-9,__consumer_offsets-46,__consumer_offsets-41,__consumer_offsets-33,__consumer_offsets-23,__consumer_offsets-49,__consumer_offsets-47,__consumer_offsets-16,__consumer_offsets-28,__consumer_offsets-31,__consumer_offsets-36,__consumer_offsets-42,__consumer_offsets-3,__consumer_offsets-18,__consumer_offsets-15,__consumer_offsets-24,__consumer_offsets-17,__consumer_offsets-48,__consumer_offsets-19,__consumer_offsets-11,__consumer_offsets-2,__consumer_offsets-43,__consumer_offsets-6,__consumer_offsets-14,__consumer_offsets-20,__consumer_offsets-0,__consumer_offsets-44,__consumer_offsets-39,__consumer_offsets-12,__consumer_offsets-45,__consumer_offsets-1,__consumer_offsets-5,__consumer_offsets-26,__consumer_offsets-29,__consumer_offsets-34,__consumer_offsets-10,__consumer_offsets-32,__consumer_offsets-40 (kafka.controller.KafkaController) -[2023-11-03 19:38:56,520] INFO [Controller id=0] New topics: [Set(test-topic)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(test-topic,Some(Pb9zfnlKRkmTGaMQyAABkw),Map(test-topic-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController) -[2023-11-03 19:38:56,520] INFO [Controller id=0] New partition creation callback for test-topic-0 (kafka.controller.KafkaController) -[2023-11-03 19:40:48,862] INFO [Controller id=0] Shutting down broker 0 (kafka.controller.KafkaController) -[2023-11-03 19:40:48,862] DEBUG [Controller id=0] All shutting down brokers: 0 (kafka.controller.KafkaController) -[2023-11-03 19:40:48,862] DEBUG [Controller id=0] Live brokers: (kafka.controller.KafkaController) -[2023-11-03 19:40:48,864] TRACE [Controller id=0] All leaders = __consumer_offsets-13 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-46 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-9 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-42 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-21 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-17 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-30 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-26 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-5 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-38 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-1 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-34 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-16 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-45 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-12 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-41 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-24 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-20 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-49 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-0 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-29 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-25 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-8 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-37 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-4 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-33 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-15 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-48 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-11 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-44 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-23 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-19 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-32 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),test-topic-0 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-28 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-7 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-40 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-3 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-36 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-47 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-14 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-43 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-10 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-22 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-18 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-31 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-27 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-39 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-6 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-35 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1),__consumer_offsets-2 -> (Leader:0,ISR:0,LeaderRecoveryState:RECOVERED,LeaderEpoch:0,ZkVersion:0,ControllerEpoch:1) (kafka.controller.KafkaController) -[2023-11-03 19:40:48,920] INFO [ControllerEventThread controllerId=0] Shutting down (kafka.controller.ControllerEventManager$ControllerEventThread) -[2023-11-03 19:40:48,921] INFO [ControllerEventThread controllerId=0] Stopped (kafka.controller.ControllerEventManager$ControllerEventThread) -[2023-11-03 19:40:48,921] INFO [ControllerEventThread controllerId=0] Shutdown completed (kafka.controller.ControllerEventManager$ControllerEventThread) -[2023-11-03 19:40:48,921] DEBUG [Controller id=0] Resigning (kafka.controller.KafkaController) -[2023-11-03 19:40:48,921] DEBUG [Controller id=0] Unregister BrokerModifications handler for Set(0) (kafka.controller.KafkaController) -[2023-11-03 19:40:48,922] INFO [PartitionStateMachine controllerId=0] Stopped partition state machine (kafka.controller.ZkPartitionStateMachine) -[2023-11-03 19:40:48,923] INFO [ReplicaStateMachine controllerId=0] Stopped replica state machine (kafka.controller.ZkReplicaStateMachine) -[2023-11-03 19:40:48,923] INFO [RequestSendThread controllerId=0] Shutting down (kafka.controller.RequestSendThread) -[2023-11-03 19:40:48,923] INFO [RequestSendThread controllerId=0] Shutdown completed (kafka.controller.RequestSendThread) -[2023-11-03 19:40:48,923] INFO [RequestSendThread controllerId=0] Stopped (kafka.controller.RequestSendThread) -[2023-11-03 19:40:48,924] INFO [Controller id=0] Resigned (kafka.controller.KafkaController) -[2023-11-03 19:45:29,427] INFO [ControllerEventThread controllerId=0] Starting (kafka.controller.ControllerEventManager$ControllerEventThread) -[2023-11-03 19:45:29,441] INFO [Controller id=0] 0 successfully elected as the controller. Epoch incremented to 2 and epoch zk version is now 2 (kafka.controller.KafkaController) -[2023-11-03 19:45:29,445] INFO [Controller id=0] Registering handlers (kafka.controller.KafkaController) -[2023-11-03 19:45:29,449] INFO [Controller id=0] Deleting log dir event notifications (kafka.controller.KafkaController) -[2023-11-03 19:45:29,451] INFO [Controller id=0] Deleting isr change notifications (kafka.controller.KafkaController) -[2023-11-03 19:45:29,453] INFO [Controller id=0] Initializing controller context (kafka.controller.KafkaController) -[2023-11-03 19:45:29,464] INFO [Controller id=0] Initialized broker epochs cache: HashMap(0 -> 156) (kafka.controller.KafkaController) -[2023-11-03 19:45:29,476] DEBUG [Controller id=0] Register BrokerModifications handler for Set(0) (kafka.controller.KafkaController) -[2023-11-03 19:45:29,509] DEBUG [Channel manager on controller 0]: Controller 0 trying to connect to broker 0 (kafka.controller.ControllerChannelManager) -[2023-11-03 19:45:29,514] INFO [RequestSendThread controllerId=0] Starting (kafka.controller.RequestSendThread) -[2023-11-03 19:45:29,515] INFO [Controller id=0] Currently active brokers in the cluster: Set(0) (kafka.controller.KafkaController) -[2023-11-03 19:45:29,515] INFO [Controller id=0] Currently shutting brokers in the cluster: HashSet() (kafka.controller.KafkaController) -[2023-11-03 19:45:29,515] INFO [Controller id=0] Current list of topics in the cluster: HashSet(test-topic, __consumer_offsets) (kafka.controller.KafkaController) -[2023-11-03 19:45:29,516] INFO [Controller id=0] Fetching topic deletions in progress (kafka.controller.KafkaController) -[2023-11-03 19:45:29,521] INFO [Controller id=0] List of topics to be deleted: (kafka.controller.KafkaController) -[2023-11-03 19:45:29,521] INFO [Controller id=0] List of topics ineligible for deletion: (kafka.controller.KafkaController) -[2023-11-03 19:45:29,521] INFO [Controller id=0] Initializing topic deletion manager (kafka.controller.KafkaController) -[2023-11-03 19:45:29,521] INFO [Topic Deletion Manager 0] Initializing manager with initial deletions: Set(), initial ineligible deletions: HashSet() (kafka.controller.TopicDeletionManager) -[2023-11-03 19:45:29,522] INFO [Controller id=0] Sending update metadata request (kafka.controller.KafkaController) -[2023-11-03 19:45:29,532] INFO [ReplicaStateMachine controllerId=0] Initializing replica state (kafka.controller.ZkReplicaStateMachine) -[2023-11-03 19:45:29,537] INFO [ReplicaStateMachine controllerId=0] Triggering online replica state changes (kafka.controller.ZkReplicaStateMachine) -[2023-11-03 19:45:29,539] WARN [RequestSendThread controllerId=0] Controller 0's connection to broker ThinkPadP53:9092 (id: 0 rack: null) was unsuccessful (kafka.controller.RequestSendThread) -java.io.IOException: Connection to ThinkPadP53:9092 (id: 0 rack: null) failed. - at org.apache.kafka.clients.NetworkClientUtils.awaitReady(NetworkClientUtils.java:70) - at kafka.controller.RequestSendThread.brokerReady(ControllerChannelManager.scala:298) - at kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:251) - at org.apache.kafka.server.util.ShutdownableThread.run(ShutdownableThread.java:130) -[2023-11-03 19:45:29,565] INFO [ReplicaStateMachine controllerId=0] Triggering offline replica state changes (kafka.controller.ZkReplicaStateMachine) -[2023-11-03 19:45:29,565] DEBUG [ReplicaStateMachine controllerId=0] Started replica state machine with initial state -> HashMap([Topic=__consumer_offsets,Partition=40,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=27,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=49,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=47,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=3,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=18,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=44,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=8,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=34,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=25,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=14,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=24,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=36,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=42,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=45,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=11,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=32,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=12,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=30,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=9,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=39,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=38,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=23,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=19,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=17,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=41,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=37,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=48,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=29,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=10,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=46,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=1,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=16,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=5,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=15,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=4,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=6,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=7,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=43,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=0,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=20,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=31,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=28,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=26,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=2,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=33,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=22,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=21,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=13,Replica=0] -> OnlineReplica, [Topic=test-topic,Partition=0,Replica=0] -> OnlineReplica, [Topic=__consumer_offsets,Partition=35,Replica=0] -> OnlineReplica) (kafka.controller.ZkReplicaStateMachine) -[2023-11-03 19:45:29,567] INFO [PartitionStateMachine controllerId=0] Initializing partition state (kafka.controller.ZkPartitionStateMachine) -[2023-11-03 19:45:29,570] INFO [PartitionStateMachine controllerId=0] Triggering online partition state changes (kafka.controller.ZkPartitionStateMachine) -[2023-11-03 19:45:29,572] DEBUG [PartitionStateMachine controllerId=0] Started partition state machine with initial state -> HashMap(__consumer_offsets-13 -> OnlinePartition, __consumer_offsets-46 -> OnlinePartition, __consumer_offsets-9 -> OnlinePartition, __consumer_offsets-42 -> OnlinePartition, __consumer_offsets-21 -> OnlinePartition, __consumer_offsets-17 -> OnlinePartition, __consumer_offsets-30 -> OnlinePartition, __consumer_offsets-26 -> OnlinePartition, __consumer_offsets-5 -> OnlinePartition, __consumer_offsets-38 -> OnlinePartition, __consumer_offsets-1 -> OnlinePartition, __consumer_offsets-34 -> OnlinePartition, __consumer_offsets-16 -> OnlinePartition, __consumer_offsets-45 -> OnlinePartition, __consumer_offsets-12 -> OnlinePartition, __consumer_offsets-41 -> OnlinePartition, __consumer_offsets-24 -> OnlinePartition, __consumer_offsets-20 -> OnlinePartition, __consumer_offsets-49 -> OnlinePartition, __consumer_offsets-0 -> OnlinePartition, __consumer_offsets-29 -> OnlinePartition, __consumer_offsets-25 -> OnlinePartition, __consumer_offsets-8 -> OnlinePartition, __consumer_offsets-37 -> OnlinePartition, __consumer_offsets-4 -> OnlinePartition, __consumer_offsets-33 -> OnlinePartition, __consumer_offsets-15 -> OnlinePartition, __consumer_offsets-48 -> OnlinePartition, __consumer_offsets-11 -> OnlinePartition, __consumer_offsets-44 -> OnlinePartition, __consumer_offsets-23 -> OnlinePartition, __consumer_offsets-19 -> OnlinePartition, __consumer_offsets-32 -> OnlinePartition, test-topic-0 -> OnlinePartition, __consumer_offsets-28 -> OnlinePartition, __consumer_offsets-7 -> OnlinePartition, __consumer_offsets-40 -> OnlinePartition, __consumer_offsets-3 -> OnlinePartition, __consumer_offsets-36 -> OnlinePartition, __consumer_offsets-47 -> OnlinePartition, __consumer_offsets-14 -> OnlinePartition, __consumer_offsets-43 -> OnlinePartition, __consumer_offsets-10 -> OnlinePartition, __consumer_offsets-22 -> OnlinePartition, __consumer_offsets-18 -> OnlinePartition, __consumer_offsets-31 -> OnlinePartition, __consumer_offsets-27 -> OnlinePartition, __consumer_offsets-39 -> OnlinePartition, __consumer_offsets-6 -> OnlinePartition, __consumer_offsets-35 -> OnlinePartition, __consumer_offsets-2 -> OnlinePartition) (kafka.controller.ZkPartitionStateMachine) -[2023-11-03 19:45:29,572] INFO [Controller id=0] Ready to serve as the new controller with epoch 2 (kafka.controller.KafkaController) -[2023-11-03 19:45:29,579] INFO [Controller id=0] Partitions undergoing preferred replica election: (kafka.controller.KafkaController) -[2023-11-03 19:45:29,579] INFO [Controller id=0] Partitions that completed preferred replica election: (kafka.controller.KafkaController) -[2023-11-03 19:45:29,579] INFO [Controller id=0] Skipping preferred replica election for partitions due to topic deletion: (kafka.controller.KafkaController) -[2023-11-03 19:45:29,580] INFO [Controller id=0] Resuming preferred replica election for partitions: (kafka.controller.KafkaController) -[2023-11-03 19:45:29,581] INFO [Controller id=0] Starting replica leader election (PREFERRED) for partitions triggered by ZkTriggered (kafka.controller.KafkaController) -[2023-11-03 19:45:29,591] INFO [Controller id=0] Starting the controller scheduler (kafka.controller.KafkaController) -[2023-11-03 19:45:29,643] INFO [RequestSendThread controllerId=0] Controller 0 connected to ThinkPadP53:9092 (id: 0 rack: null) for sending state change requests (kafka.controller.RequestSendThread) -[2023-11-03 19:45:34,592] INFO [Controller id=0] Processing automatic preferred replica leader election (kafka.controller.KafkaController) -[2023-11-03 19:45:34,593] TRACE [Controller id=0] Checking need to trigger auto leader balancing (kafka.controller.KafkaController) -[2023-11-03 19:45:34,595] DEBUG [Controller id=0] Topics not in preferred replica for broker 0 HashMap() (kafka.controller.KafkaController) -[2023-11-03 19:45:34,596] TRACE [Controller id=0] Leader imbalance ratio for broker 0 is 0.0 (kafka.controller.KafkaController) +[2023-11-06 13:26:32,055] INFO [Controller id=0] Partitions undergoing preferred replica election: (kafka.controller.KafkaController) +[2023-11-06 13:26:32,055] INFO [Controller id=0] Partitions that completed preferred replica election: (kafka.controller.KafkaController) +[2023-11-06 13:26:32,055] INFO [Controller id=0] Skipping preferred replica election for partitions due to topic deletion: (kafka.controller.KafkaController) +[2023-11-06 13:26:32,056] INFO [Controller id=0] Resuming preferred replica election for partitions: (kafka.controller.KafkaController) +[2023-11-06 13:26:32,056] INFO [Controller id=0] Starting replica leader election (PREFERRED) for partitions triggered by ZkTriggered (kafka.controller.KafkaController) +[2023-11-06 13:26:32,063] INFO [Controller id=0] Starting the controller scheduler (kafka.controller.KafkaController) +[2023-11-06 13:26:32,156] INFO [RequestSendThread controllerId=0] Controller 0 connected to localhost:9092 (id: 0 rack: null) for sending state change requests (kafka.controller.RequestSendThread) +[2023-11-06 13:26:37,064] INFO [Controller id=0] Processing automatic preferred replica leader election (kafka.controller.KafkaController) +[2023-11-06 13:26:37,064] TRACE [Controller id=0] Checking need to trigger auto leader balancing (kafka.controller.KafkaController) +[2023-11-06 13:27:16,510] INFO [Controller id=0] New topics: [Set(__consumer_offsets)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(__consumer_offsets,Some(f9d_z6FzSde58txrT_Qj9w),HashMap(__consumer_offsets-22 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-30 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-25 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-35 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-37 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-38 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-13 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-8 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-21 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-4 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-27 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-7 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-9 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-46 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-41 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-33 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-23 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-49 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-47 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-16 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-28 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-31 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-36 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-42 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-3 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-18 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-15 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-24 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-17 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-48 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-19 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-11 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-2 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-43 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-6 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-14 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-20 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-44 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-39 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-12 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-45 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-1 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-5 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-26 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-29 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-34 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-10 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-32 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=), __consumer_offsets-40 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController) +[2023-11-06 13:27:16,511] INFO [Controller id=0] New partition creation callback for __consumer_offsets-22,__consumer_offsets-30,__consumer_offsets-25,__consumer_offsets-35,__consumer_offsets-37,__consumer_offsets-38,__consumer_offsets-13,__consumer_offsets-8,__consumer_offsets-21,__consumer_offsets-4,__consumer_offsets-27,__consumer_offsets-7,__consumer_offsets-9,__consumer_offsets-46,__consumer_offsets-41,__consumer_offsets-33,__consumer_offsets-23,__consumer_offsets-49,__consumer_offsets-47,__consumer_offsets-16,__consumer_offsets-28,__consumer_offsets-31,__consumer_offsets-36,__consumer_offsets-42,__consumer_offsets-3,__consumer_offsets-18,__consumer_offsets-15,__consumer_offsets-24,__consumer_offsets-17,__consumer_offsets-48,__consumer_offsets-19,__consumer_offsets-11,__consumer_offsets-2,__consumer_offsets-43,__consumer_offsets-6,__consumer_offsets-14,__consumer_offsets-20,__consumer_offsets-0,__consumer_offsets-44,__consumer_offsets-39,__consumer_offsets-12,__consumer_offsets-45,__consumer_offsets-1,__consumer_offsets-5,__consumer_offsets-26,__consumer_offsets-29,__consumer_offsets-34,__consumer_offsets-10,__consumer_offsets-32,__consumer_offsets-40 (kafka.controller.KafkaController) +[2023-11-06 13:27:32,128] INFO [Controller id=0] New topics: [Set(test-topic)], deleted topics: [HashSet()], new partition replica assignment [Set(TopicIdReplicaAssignment(test-topic,Some(Hx76FWANRJGp_-YQs8849Q),Map(test-topic-0 -> ReplicaAssignment(replicas=0, addingReplicas=, removingReplicas=))))] (kafka.controller.KafkaController) +[2023-11-06 13:27:32,128] INFO [Controller id=0] New partition creation callback for test-topic-0 (kafka.controller.KafkaController) diff --git a/logs/kafkaServer-gc.log b/logs/kafkaServer-gc.log index 17b30bc..c4bc3ac 100644 --- a/logs/kafkaServer-gc.log +++ b/logs/kafkaServer-gc.log @@ -1,125 +1,125 @@ -[2023-11-03T19:45:27.999-0400][gc] Using G1 -[2023-11-03T19:45:28.005-0400][gc,init] Version: 17.0.6+10 (release) -[2023-11-03T19:45:28.005-0400][gc,init] CPUs: 12 total, 12 available -[2023-11-03T19:45:28.005-0400][gc,init] Memory: 63941M -[2023-11-03T19:45:28.005-0400][gc,init] Large Page Support: Disabled -[2023-11-03T19:45:28.005-0400][gc,init] NUMA Support: Disabled -[2023-11-03T19:45:28.005-0400][gc,init] Compressed Oops: Enabled (32-bit) -[2023-11-03T19:45:28.005-0400][gc,init] Heap Region Size: 1M -[2023-11-03T19:45:28.005-0400][gc,init] Heap Min Capacity: 1G -[2023-11-03T19:45:28.005-0400][gc,init] Heap Initial Capacity: 1G -[2023-11-03T19:45:28.005-0400][gc,init] Heap Max Capacity: 1G -[2023-11-03T19:45:28.005-0400][gc,init] Pre-touch: Disabled -[2023-11-03T19:45:28.005-0400][gc,init] Parallel Workers: 10 -[2023-11-03T19:45:28.005-0400][gc,init] Concurrent Workers: 3 -[2023-11-03T19:45:28.005-0400][gc,init] Concurrent Refinement Workers: 10 -[2023-11-03T19:45:28.005-0400][gc,init] Periodic GC: Disabled -[2023-11-03T19:45:28.005-0400][gc,metaspace] CDS archive(s) mapped at: [0x0000000800000000-0x0000000800bd5000-0x0000000800bd5000), size 12406784, SharedBaseAddress: 0x0000000800000000, ArchiveRelocationMode: 0. -[2023-11-03T19:45:28.005-0400][gc,metaspace] Compressed class space mapped at: 0x0000000800c00000-0x0000000840c00000, reserved size: 1073741824 -[2023-11-03T19:45:28.005-0400][gc,metaspace] Narrow klass base: 0x0000000800000000, Narrow klass shift: 0, Narrow klass range: 0x100000000 -[2023-11-03T19:45:28.714-0400][gc,start ] GC(0) Pause Young (Normal) (G1 Evacuation Pause) -[2023-11-03T19:45:28.715-0400][gc,task ] GC(0) Using 10 workers of 10 for evacuation -[2023-11-03T19:45:28.722-0400][gc,phases ] GC(0) Pre Evacuate Collection Set: 0.1ms -[2023-11-03T19:45:28.722-0400][gc,phases ] GC(0) Merge Heap Roots: 0.1ms -[2023-11-03T19:45:28.722-0400][gc,phases ] GC(0) Evacuate Collection Set: 6.5ms -[2023-11-03T19:45:28.722-0400][gc,phases ] GC(0) Post Evacuate Collection Set: 0.4ms -[2023-11-03T19:45:28.722-0400][gc,phases ] GC(0) Other: 0.8ms -[2023-11-03T19:45:28.722-0400][gc,heap ] GC(0) Eden regions: 51->0(44) -[2023-11-03T19:45:28.722-0400][gc,heap ] GC(0) Survivor regions: 0->7(7) -[2023-11-03T19:45:28.722-0400][gc,heap ] GC(0) Old regions: 0->1 -[2023-11-03T19:45:28.722-0400][gc,heap ] GC(0) Archive regions: 2->2 -[2023-11-03T19:45:28.722-0400][gc,heap ] GC(0) Humongous regions: 0->0 -[2023-11-03T19:45:28.722-0400][gc,metaspace] GC(0) Metaspace: 18659K(18880K)->18659K(18880K) NonClass: 16455K(16576K)->16455K(16576K) Class: 2203K(2304K)->2203K(2304K) -[2023-11-03T19:45:28.722-0400][gc ] GC(0) Pause Young (Normal) (G1 Evacuation Pause) 51M->8M(1024M) 7.973ms -[2023-11-03T19:45:28.722-0400][gc,cpu ] GC(0) User=0.07s Sys=0.00s Real=0.01s -[2023-11-03T19:45:28.862-0400][gc,start ] GC(1) Pause Young (Concurrent Start) (Metadata GC Threshold) -[2023-11-03T19:45:28.862-0400][gc,task ] GC(1) Using 10 workers of 10 for evacuation -[2023-11-03T19:45:28.869-0400][gc,phases ] GC(1) Pre Evacuate Collection Set: 0.1ms -[2023-11-03T19:45:28.869-0400][gc,phases ] GC(1) Merge Heap Roots: 0.0ms -[2023-11-03T19:45:28.869-0400][gc,phases ] GC(1) Evacuate Collection Set: 6.4ms -[2023-11-03T19:45:28.869-0400][gc,phases ] GC(1) Post Evacuate Collection Set: 0.4ms -[2023-11-03T19:45:28.869-0400][gc,phases ] GC(1) Other: 0.2ms -[2023-11-03T19:45:28.869-0400][gc,heap ] GC(1) Eden regions: 10->0(50) -[2023-11-03T19:45:28.869-0400][gc,heap ] GC(1) Survivor regions: 7->1(7) -[2023-11-03T19:45:28.869-0400][gc,heap ] GC(1) Old regions: 1->8 -[2023-11-03T19:45:28.869-0400][gc,heap ] GC(1) Archive regions: 2->2 -[2023-11-03T19:45:28.869-0400][gc,heap ] GC(1) Humongous regions: 0->0 -[2023-11-03T19:45:28.869-0400][gc,metaspace] GC(1) Metaspace: 21310K(21504K)->21310K(21504K) NonClass: 18734K(18816K)->18734K(18816K) Class: 2575K(2688K)->2575K(2688K) -[2023-11-03T19:45:28.869-0400][gc ] GC(1) Pause Young (Concurrent Start) (Metadata GC Threshold) 18M->9M(1024M) 7.138ms -[2023-11-03T19:45:28.869-0400][gc,cpu ] GC(1) User=0.00s Sys=0.06s Real=0.01s -[2023-11-03T19:45:28.869-0400][gc ] GC(2) Concurrent Mark Cycle -[2023-11-03T19:45:28.869-0400][gc,marking ] GC(2) Concurrent Clear Claimed Marks -[2023-11-03T19:45:28.869-0400][gc,marking ] GC(2) Concurrent Clear Claimed Marks 0.011ms -[2023-11-03T19:45:28.869-0400][gc,marking ] GC(2) Concurrent Scan Root Regions -[2023-11-03T19:45:28.871-0400][gc,marking ] GC(2) Concurrent Scan Root Regions 1.137ms -[2023-11-03T19:45:28.871-0400][gc,marking ] GC(2) Concurrent Mark -[2023-11-03T19:45:28.871-0400][gc,marking ] GC(2) Concurrent Mark From Roots -[2023-11-03T19:45:28.871-0400][gc,task ] GC(2) Using 3 workers of 3 for marking -[2023-11-03T19:45:28.871-0400][gc,marking ] GC(2) Concurrent Mark From Roots 0.977ms -[2023-11-03T19:45:28.872-0400][gc,marking ] GC(2) Concurrent Preclean -[2023-11-03T19:45:28.872-0400][gc,marking ] GC(2) Concurrent Preclean 0.054ms -[2023-11-03T19:45:28.872-0400][gc,start ] GC(2) Pause Remark -[2023-11-03T19:45:28.873-0400][gc ] GC(2) Pause Remark 10M->10M(1024M) 0.908ms -[2023-11-03T19:45:28.873-0400][gc,cpu ] GC(2) User=0.00s Sys=0.00s Real=0.00s -[2023-11-03T19:45:28.873-0400][gc,marking ] GC(2) Concurrent Mark 2.112ms -[2023-11-03T19:45:28.873-0400][gc,marking ] GC(2) Concurrent Rebuild Remembered Sets -[2023-11-03T19:45:28.874-0400][gc,marking ] GC(2) Concurrent Rebuild Remembered Sets 1.239ms -[2023-11-03T19:45:28.874-0400][gc,start ] GC(2) Pause Cleanup -[2023-11-03T19:45:28.874-0400][gc ] GC(2) Pause Cleanup 10M->10M(1024M) 0.183ms -[2023-11-03T19:45:28.874-0400][gc,cpu ] GC(2) User=0.00s Sys=0.00s Real=0.00s -[2023-11-03T19:45:28.874-0400][gc,marking ] GC(2) Concurrent Cleanup for Next Mark -[2023-11-03T19:45:28.878-0400][gc,marking ] GC(2) Concurrent Cleanup for Next Mark 3.944ms -[2023-11-03T19:45:28.878-0400][gc ] GC(2) Concurrent Mark Cycle 8.835ms -[2023-11-03T19:45:29.405-0400][gc,start ] GC(3) Pause Young (Normal) (G1 Evacuation Pause) -[2023-11-03T19:45:29.405-0400][gc,task ] GC(3) Using 10 workers of 10 for evacuation -[2023-11-03T19:45:29.410-0400][gc,phases ] GC(3) Pre Evacuate Collection Set: 0.1ms -[2023-11-03T19:45:29.410-0400][gc,phases ] GC(3) Merge Heap Roots: 0.1ms -[2023-11-03T19:45:29.410-0400][gc,phases ] GC(3) Evacuate Collection Set: 2.7ms -[2023-11-03T19:45:29.410-0400][gc,phases ] GC(3) Post Evacuate Collection Set: 1.7ms -[2023-11-03T19:45:29.410-0400][gc,phases ] GC(3) Other: 0.2ms -[2023-11-03T19:45:29.410-0400][gc,heap ] GC(3) Eden regions: 50->0(45) -[2023-11-03T19:45:29.410-0400][gc,heap ] GC(3) Survivor regions: 1->6(7) -[2023-11-03T19:45:29.410-0400][gc,heap ] GC(3) Old regions: 8->8 -[2023-11-03T19:45:29.410-0400][gc,heap ] GC(3) Archive regions: 2->2 -[2023-11-03T19:45:29.410-0400][gc,heap ] GC(3) Humongous regions: 129->129 -[2023-11-03T19:45:29.410-0400][gc,metaspace] GC(3) Metaspace: 30840K(31168K)->30840K(31168K) NonClass: 27360K(27520K)->27360K(27520K) Class: 3480K(3648K)->3480K(3648K) -[2023-11-03T19:45:29.410-0400][gc ] GC(3) Pause Young (Normal) (G1 Evacuation Pause) 188M->143M(1024M) 4.864ms -[2023-11-03T19:45:29.410-0400][gc,cpu ] GC(3) User=0.03s Sys=0.00s Real=0.00s -[2023-11-03T19:45:29.651-0400][gc,start ] GC(4) Pause Young (Concurrent Start) (Metadata GC Threshold) -[2023-11-03T19:45:29.651-0400][gc,task ] GC(4) Using 10 workers of 10 for evacuation -[2023-11-03T19:45:29.656-0400][gc,phases ] GC(4) Pre Evacuate Collection Set: 0.1ms -[2023-11-03T19:45:29.656-0400][gc,phases ] GC(4) Merge Heap Roots: 0.0ms -[2023-11-03T19:45:29.656-0400][gc,phases ] GC(4) Evacuate Collection Set: 3.8ms -[2023-11-03T19:45:29.656-0400][gc,phases ] GC(4) Post Evacuate Collection Set: 0.3ms -[2023-11-03T19:45:29.656-0400][gc,phases ] GC(4) Other: 0.1ms -[2023-11-03T19:45:29.656-0400][gc,heap ] GC(4) Eden regions: 21->0(49) -[2023-11-03T19:45:29.656-0400][gc,heap ] GC(4) Survivor regions: 6->2(7) -[2023-11-03T19:45:29.656-0400][gc,heap ] GC(4) Old regions: 8->13 -[2023-11-03T19:45:29.656-0400][gc,heap ] GC(4) Archive regions: 2->2 -[2023-11-03T19:45:29.656-0400][gc,heap ] GC(4) Humongous regions: 129->129 -[2023-11-03T19:45:29.656-0400][gc,metaspace] GC(4) Metaspace: 35698K(35968K)->35698K(35968K) NonClass: 31646K(31808K)->31646K(31808K) Class: 4051K(4160K)->4051K(4160K) -[2023-11-03T19:45:29.656-0400][gc ] GC(4) Pause Young (Concurrent Start) (Metadata GC Threshold) 163M->144M(1024M) 4.413ms -[2023-11-03T19:45:29.656-0400][gc,cpu ] GC(4) User=0.05s Sys=0.01s Real=0.01s -[2023-11-03T19:45:29.656-0400][gc ] GC(5) Concurrent Mark Cycle -[2023-11-03T19:45:29.656-0400][gc,marking ] GC(5) Concurrent Clear Claimed Marks -[2023-11-03T19:45:29.656-0400][gc,marking ] GC(5) Concurrent Clear Claimed Marks 0.017ms -[2023-11-03T19:45:29.656-0400][gc,marking ] GC(5) Concurrent Scan Root Regions -[2023-11-03T19:45:29.658-0400][gc,marking ] GC(5) Concurrent Scan Root Regions 1.818ms -[2023-11-03T19:45:29.658-0400][gc,marking ] GC(5) Concurrent Mark -[2023-11-03T19:45:29.658-0400][gc,marking ] GC(5) Concurrent Mark From Roots -[2023-11-03T19:45:29.658-0400][gc,task ] GC(5) Using 3 workers of 3 for marking -[2023-11-03T19:45:29.662-0400][gc,marking ] GC(5) Concurrent Mark From Roots 4.776ms -[2023-11-03T19:45:29.662-0400][gc,marking ] GC(5) Concurrent Preclean -[2023-11-03T19:45:29.663-0400][gc,marking ] GC(5) Concurrent Preclean 0.125ms -[2023-11-03T19:45:29.663-0400][gc,start ] GC(5) Pause Remark -[2023-11-03T19:45:29.664-0400][gc ] GC(5) Pause Remark 145M->145M(1024M) 0.870ms -[2023-11-03T19:45:29.664-0400][gc,cpu ] GC(5) User=0.01s Sys=0.00s Real=0.00s -[2023-11-03T19:45:29.664-0400][gc,marking ] GC(5) Concurrent Mark 5.944ms -[2023-11-03T19:45:29.664-0400][gc,marking ] GC(5) Concurrent Rebuild Remembered Sets -[2023-11-03T19:45:29.664-0400][gc,marking ] GC(5) Concurrent Rebuild Remembered Sets 0.004ms -[2023-11-03T19:45:29.664-0400][gc,start ] GC(5) Pause Cleanup -[2023-11-03T19:45:29.664-0400][gc ] GC(5) Pause Cleanup 145M->145M(1024M) 0.006ms -[2023-11-03T19:45:29.664-0400][gc,cpu ] GC(5) User=0.00s Sys=0.00s Real=0.00s -[2023-11-03T19:45:29.664-0400][gc,marking ] GC(5) Concurrent Cleanup for Next Mark -[2023-11-03T19:45:29.665-0400][gc,marking ] GC(5) Concurrent Cleanup for Next Mark 1.530ms -[2023-11-03T19:45:29.665-0400][gc ] GC(5) Concurrent Mark Cycle 9.512ms +[2023-11-06T13:26:30.795-0500][gc] Using G1 +[2023-11-06T13:26:30.801-0500][gc,init] Version: 17.0.8.1+1 (release) +[2023-11-06T13:26:30.801-0500][gc,init] CPUs: 12 total, 12 available +[2023-11-06T13:26:30.801-0500][gc,init] Memory: 63941M +[2023-11-06T13:26:30.801-0500][gc,init] Large Page Support: Disabled +[2023-11-06T13:26:30.801-0500][gc,init] NUMA Support: Disabled +[2023-11-06T13:26:30.801-0500][gc,init] Compressed Oops: Enabled (32-bit) +[2023-11-06T13:26:30.801-0500][gc,init] Heap Region Size: 1M +[2023-11-06T13:26:30.801-0500][gc,init] Heap Min Capacity: 1G +[2023-11-06T13:26:30.801-0500][gc,init] Heap Initial Capacity: 1G +[2023-11-06T13:26:30.801-0500][gc,init] Heap Max Capacity: 1G +[2023-11-06T13:26:30.801-0500][gc,init] Pre-touch: Disabled +[2023-11-06T13:26:30.801-0500][gc,init] Parallel Workers: 10 +[2023-11-06T13:26:30.801-0500][gc,init] Concurrent Workers: 3 +[2023-11-06T13:26:30.801-0500][gc,init] Concurrent Refinement Workers: 10 +[2023-11-06T13:26:30.801-0500][gc,init] Periodic GC: Disabled +[2023-11-06T13:26:30.805-0500][gc,metaspace] CDS archive(s) mapped at: [0x00007f77f7000000-0x00007f77f7beb000-0x00007f77f7beb000), size 12496896, SharedBaseAddress: 0x00007f77f7000000, ArchiveRelocationMode: 1. +[2023-11-06T13:26:30.805-0500][gc,metaspace] Compressed class space mapped at: 0x00007f77f8000000-0x00007f7838000000, reserved size: 1073741824 +[2023-11-06T13:26:30.805-0500][gc,metaspace] Narrow klass base: 0x00007f77f7000000, Narrow klass shift: 0, Narrow klass range: 0x100000000 +[2023-11-06T13:26:31.511-0500][gc,start ] GC(0) Pause Young (Normal) (G1 Evacuation Pause) +[2023-11-06T13:26:31.512-0500][gc,task ] GC(0) Using 10 workers of 10 for evacuation +[2023-11-06T13:26:31.518-0500][gc,phases ] GC(0) Pre Evacuate Collection Set: 0.1ms +[2023-11-06T13:26:31.518-0500][gc,phases ] GC(0) Merge Heap Roots: 0.1ms +[2023-11-06T13:26:31.518-0500][gc,phases ] GC(0) Evacuate Collection Set: 5.7ms +[2023-11-06T13:26:31.518-0500][gc,phases ] GC(0) Post Evacuate Collection Set: 0.5ms +[2023-11-06T13:26:31.518-0500][gc,phases ] GC(0) Other: 0.8ms +[2023-11-06T13:26:31.518-0500][gc,heap ] GC(0) Eden regions: 51->0(44) +[2023-11-06T13:26:31.518-0500][gc,heap ] GC(0) Survivor regions: 0->7(7) +[2023-11-06T13:26:31.518-0500][gc,heap ] GC(0) Old regions: 0->1 +[2023-11-06T13:26:31.518-0500][gc,heap ] GC(0) Archive regions: 2->2 +[2023-11-06T13:26:31.518-0500][gc,heap ] GC(0) Humongous regions: 0->0 +[2023-11-06T13:26:31.518-0500][gc,metaspace] GC(0) Metaspace: 18606K(18816K)->18606K(18816K) NonClass: 16396K(16512K)->16396K(16512K) Class: 2210K(2304K)->2210K(2304K) +[2023-11-06T13:26:31.518-0500][gc ] GC(0) Pause Young (Normal) (G1 Evacuation Pause) 51M->8M(1024M) 7.181ms +[2023-11-06T13:26:31.518-0500][gc,cpu ] GC(0) User=0.05s Sys=0.00s Real=0.00s +[2023-11-06T13:26:31.655-0500][gc,start ] GC(1) Pause Young (Concurrent Start) (Metadata GC Threshold) +[2023-11-06T13:26:31.655-0500][gc,task ] GC(1) Using 10 workers of 10 for evacuation +[2023-11-06T13:26:31.662-0500][gc,phases ] GC(1) Pre Evacuate Collection Set: 0.1ms +[2023-11-06T13:26:31.662-0500][gc,phases ] GC(1) Merge Heap Roots: 0.1ms +[2023-11-06T13:26:31.662-0500][gc,phases ] GC(1) Evacuate Collection Set: 6.6ms +[2023-11-06T13:26:31.662-0500][gc,phases ] GC(1) Post Evacuate Collection Set: 0.5ms +[2023-11-06T13:26:31.662-0500][gc,phases ] GC(1) Other: 0.2ms +[2023-11-06T13:26:31.662-0500][gc,heap ] GC(1) Eden regions: 12->0(49) +[2023-11-06T13:26:31.662-0500][gc,heap ] GC(1) Survivor regions: 7->2(7) +[2023-11-06T13:26:31.662-0500][gc,heap ] GC(1) Old regions: 1->8 +[2023-11-06T13:26:31.662-0500][gc,heap ] GC(1) Archive regions: 2->2 +[2023-11-06T13:26:31.662-0500][gc,heap ] GC(1) Humongous regions: 129->129 +[2023-11-06T13:26:31.662-0500][gc,metaspace] GC(1) Metaspace: 21299K(21504K)->21299K(21504K) NonClass: 18739K(18816K)->18739K(18816K) Class: 2559K(2688K)->2559K(2688K) +[2023-11-06T13:26:31.662-0500][gc ] GC(1) Pause Young (Concurrent Start) (Metadata GC Threshold) 149M->138M(1024M) 7.560ms +[2023-11-06T13:26:31.662-0500][gc,cpu ] GC(1) User=0.00s Sys=0.06s Real=0.01s +[2023-11-06T13:26:31.662-0500][gc ] GC(2) Concurrent Mark Cycle +[2023-11-06T13:26:31.662-0500][gc,marking ] GC(2) Concurrent Clear Claimed Marks +[2023-11-06T13:26:31.662-0500][gc,marking ] GC(2) Concurrent Clear Claimed Marks 0.011ms +[2023-11-06T13:26:31.662-0500][gc,marking ] GC(2) Concurrent Scan Root Regions +[2023-11-06T13:26:31.664-0500][gc,marking ] GC(2) Concurrent Scan Root Regions 1.308ms +[2023-11-06T13:26:31.664-0500][gc,marking ] GC(2) Concurrent Mark +[2023-11-06T13:26:31.664-0500][gc,marking ] GC(2) Concurrent Mark From Roots +[2023-11-06T13:26:31.664-0500][gc,task ] GC(2) Using 3 workers of 3 for marking +[2023-11-06T13:26:31.665-0500][gc,marking ] GC(2) Concurrent Mark From Roots 1.090ms +[2023-11-06T13:26:31.665-0500][gc,marking ] GC(2) Concurrent Preclean +[2023-11-06T13:26:31.665-0500][gc,marking ] GC(2) Concurrent Preclean 0.055ms +[2023-11-06T13:26:31.665-0500][gc,start ] GC(2) Pause Remark +[2023-11-06T13:26:31.666-0500][gc ] GC(2) Pause Remark 139M->139M(1024M) 1.293ms +[2023-11-06T13:26:31.666-0500][gc,cpu ] GC(2) User=0.00s Sys=0.01s Real=0.00s +[2023-11-06T13:26:31.666-0500][gc,marking ] GC(2) Concurrent Mark 2.671ms +[2023-11-06T13:26:31.666-0500][gc,marking ] GC(2) Concurrent Rebuild Remembered Sets +[2023-11-06T13:26:31.668-0500][gc,marking ] GC(2) Concurrent Rebuild Remembered Sets 1.262ms +[2023-11-06T13:26:31.668-0500][gc,start ] GC(2) Pause Cleanup +[2023-11-06T13:26:31.668-0500][gc ] GC(2) Pause Cleanup 139M->139M(1024M) 0.181ms +[2023-11-06T13:26:31.668-0500][gc,cpu ] GC(2) User=0.00s Sys=0.00s Real=0.00s +[2023-11-06T13:26:31.668-0500][gc,marking ] GC(2) Concurrent Cleanup for Next Mark +[2023-11-06T13:26:31.669-0500][gc,marking ] GC(2) Concurrent Cleanup for Next Mark 1.442ms +[2023-11-06T13:26:31.669-0500][gc ] GC(2) Concurrent Mark Cycle 7.132ms +[2023-11-06T13:26:32.046-0500][gc,start ] GC(3) Pause Young (Normal) (G1 Evacuation Pause) +[2023-11-06T13:26:32.046-0500][gc,task ] GC(3) Using 10 workers of 10 for evacuation +[2023-11-06T13:26:32.048-0500][gc,phases ] GC(3) Pre Evacuate Collection Set: 0.1ms +[2023-11-06T13:26:32.048-0500][gc,phases ] GC(3) Merge Heap Roots: 0.1ms +[2023-11-06T13:26:32.048-0500][gc,phases ] GC(3) Evacuate Collection Set: 1.6ms +[2023-11-06T13:26:32.048-0500][gc,phases ] GC(3) Post Evacuate Collection Set: 0.3ms +[2023-11-06T13:26:32.048-0500][gc,phases ] GC(3) Other: 0.1ms +[2023-11-06T13:26:32.048-0500][gc,heap ] GC(3) Eden regions: 49->0(45) +[2023-11-06T13:26:32.048-0500][gc,heap ] GC(3) Survivor regions: 2->6(7) +[2023-11-06T13:26:32.048-0500][gc,heap ] GC(3) Old regions: 8->8 +[2023-11-06T13:26:32.048-0500][gc,heap ] GC(3) Archive regions: 2->2 +[2023-11-06T13:26:32.048-0500][gc,heap ] GC(3) Humongous regions: 129->129 +[2023-11-06T13:26:32.048-0500][gc,metaspace] GC(3) Metaspace: 32579K(32832K)->32579K(32832K) NonClass: 28947K(29120K)->28947K(29120K) Class: 3631K(3712K)->3631K(3712K) +[2023-11-06T13:26:32.048-0500][gc ] GC(3) Pause Young (Normal) (G1 Evacuation Pause) 187M->143M(1024M) 2.247ms +[2023-11-06T13:26:32.048-0500][gc,cpu ] GC(3) User=0.02s Sys=0.00s Real=0.00s +[2023-11-06T13:27:16.575-0500][gc,start ] GC(4) Pause Young (Concurrent Start) (Metadata GC Threshold) +[2023-11-06T13:27:16.575-0500][gc,task ] GC(4) Using 10 workers of 10 for evacuation +[2023-11-06T13:27:16.580-0500][gc,phases ] GC(4) Pre Evacuate Collection Set: 0.1ms +[2023-11-06T13:27:16.580-0500][gc,phases ] GC(4) Merge Heap Roots: 0.0ms +[2023-11-06T13:27:16.580-0500][gc,phases ] GC(4) Evacuate Collection Set: 4.0ms +[2023-11-06T13:27:16.580-0500][gc,phases ] GC(4) Post Evacuate Collection Set: 0.4ms +[2023-11-06T13:27:16.580-0500][gc,phases ] GC(4) Other: 0.1ms +[2023-11-06T13:27:16.580-0500][gc,heap ] GC(4) Eden regions: 17->0(49) +[2023-11-06T13:27:16.580-0500][gc,heap ] GC(4) Survivor regions: 6->2(7) +[2023-11-06T13:27:16.580-0500][gc,heap ] GC(4) Old regions: 8->14 +[2023-11-06T13:27:16.580-0500][gc,heap ] GC(4) Archive regions: 2->2 +[2023-11-06T13:27:16.580-0500][gc,heap ] GC(4) Humongous regions: 129->129 +[2023-11-06T13:27:16.580-0500][gc,metaspace] GC(4) Metaspace: 35667K(35968K)->35667K(35968K) NonClass: 31529K(31680K)->31529K(31680K) Class: 4138K(4288K)->4138K(4288K) +[2023-11-06T13:27:16.580-0500][gc ] GC(4) Pause Young (Concurrent Start) (Metadata GC Threshold) 159M->144M(1024M) 4.693ms +[2023-11-06T13:27:16.580-0500][gc,cpu ] GC(4) User=0.04s Sys=0.01s Real=0.01s +[2023-11-06T13:27:16.580-0500][gc ] GC(5) Concurrent Mark Cycle +[2023-11-06T13:27:16.580-0500][gc,marking ] GC(5) Concurrent Clear Claimed Marks +[2023-11-06T13:27:16.580-0500][gc,marking ] GC(5) Concurrent Clear Claimed Marks 0.047ms +[2023-11-06T13:27:16.580-0500][gc,marking ] GC(5) Concurrent Scan Root Regions +[2023-11-06T13:27:16.582-0500][gc,marking ] GC(5) Concurrent Scan Root Regions 1.685ms +[2023-11-06T13:27:16.582-0500][gc,marking ] GC(5) Concurrent Mark +[2023-11-06T13:27:16.582-0500][gc,marking ] GC(5) Concurrent Mark From Roots +[2023-11-06T13:27:16.582-0500][gc,task ] GC(5) Using 3 workers of 3 for marking +[2023-11-06T13:27:16.585-0500][gc,marking ] GC(5) Concurrent Mark From Roots 3.066ms +[2023-11-06T13:27:16.585-0500][gc,marking ] GC(5) Concurrent Preclean +[2023-11-06T13:27:16.585-0500][gc,marking ] GC(5) Concurrent Preclean 0.080ms +[2023-11-06T13:27:16.585-0500][gc,start ] GC(5) Pause Remark +[2023-11-06T13:27:16.586-0500][gc ] GC(5) Pause Remark 144M->144M(1024M) 0.978ms +[2023-11-06T13:27:16.586-0500][gc,cpu ] GC(5) User=0.01s Sys=0.00s Real=0.00s +[2023-11-06T13:27:16.586-0500][gc,marking ] GC(5) Concurrent Mark 4.710ms +[2023-11-06T13:27:16.586-0500][gc,marking ] GC(5) Concurrent Rebuild Remembered Sets +[2023-11-06T13:27:16.589-0500][gc,marking ] GC(5) Concurrent Rebuild Remembered Sets 2.484ms +[2023-11-06T13:27:16.589-0500][gc,start ] GC(5) Pause Cleanup +[2023-11-06T13:27:16.589-0500][gc ] GC(5) Pause Cleanup 145M->145M(1024M) 0.124ms +[2023-11-06T13:27:16.589-0500][gc,cpu ] GC(5) User=0.00s Sys=0.00s Real=0.00s +[2023-11-06T13:27:16.589-0500][gc,marking ] GC(5) Concurrent Cleanup for Next Mark +[2023-11-06T13:27:16.592-0500][gc,marking ] GC(5) Concurrent Cleanup for Next Mark 2.206ms +[2023-11-06T13:27:16.592-0500][gc ] GC(5) Concurrent Mark Cycle 11.834ms diff --git a/logs/log-cleaner.log b/logs/log-cleaner.log index 36119a2..75ad872 100644 --- a/logs/log-cleaner.log +++ b/logs/log-cleaner.log @@ -1,3 +1 @@ -[2023-11-03 19:38:20,717] INFO Starting the log cleaner (kafka.log.LogCleaner) -[2023-11-03 19:40:48,894] INFO Shutting down the log cleaner. (kafka.log.LogCleaner) -[2023-11-03 19:45:29,067] INFO Starting the log cleaner (kafka.log.LogCleaner) +[2023-11-06 13:26:31,617] INFO Starting the log cleaner (kafka.log.LogCleaner) diff --git a/logs/server.log b/logs/server.log index 04f6372..2ed5810 100644 --- a/logs/server.log +++ b/logs/server.log @@ -1,127 +1,127 @@ -[2023-11-03 19:38:10,839] INFO Reading configuration from: ./config/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:38:10,841] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:38:10,841] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:38:10,841] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:38:10,841] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:38:10,844] INFO autopurge.snapRetainCount set to 3 (org.apache.zookeeper.server.DatadirCleanupManager) -[2023-11-03 19:38:10,844] INFO autopurge.purgeInterval set to 0 (org.apache.zookeeper.server.DatadirCleanupManager) -[2023-11-03 19:38:10,844] INFO Purge task is not scheduled. (org.apache.zookeeper.server.DatadirCleanupManager) -[2023-11-03 19:38:10,844] WARN Either no config or no quorum defined in config, running in standalone mode (org.apache.zookeeper.server.quorum.QuorumPeerMain) -[2023-11-03 19:38:10,845] INFO Log4j 1.2 jmx support not found; jmx disabled. (org.apache.zookeeper.jmx.ManagedUtil) -[2023-11-03 19:38:10,845] INFO Reading configuration from: ./config/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:38:10,846] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:38:10,846] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:38:10,846] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:38:10,846] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:38:10,846] INFO Starting server (org.apache.zookeeper.server.ZooKeeperServerMain) -[2023-11-03 19:38:10,854] INFO ServerMetrics initialized with provider org.apache.zookeeper.metrics.impl.DefaultMetricsProvider@e50a6f6 (org.apache.zookeeper.server.ServerMetrics) -[2023-11-03 19:38:10,857] INFO ACL digest algorithm is: SHA1 (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) -[2023-11-03 19:38:10,857] INFO zookeeper.DigestAuthenticationProvider.enabled = true (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) -[2023-11-03 19:38:10,861] INFO zookeeper.snapshot.trust.empty : false (org.apache.zookeeper.server.persistence.FileTxnSnapLog) -[2023-11-03 19:38:10,868] INFO (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,868] INFO ______ _ (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,868] INFO |___ / | | (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,868] INFO / / ___ ___ | | __ ___ ___ _ __ ___ _ __ (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,868] INFO / / / _ \ / _ \ | |/ / / _ \ / _ \ | '_ \ / _ \ | '__| (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,868] INFO / /__ | (_) | | (_) | | < | __/ | __/ | |_) | | __/ | | (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,868] INFO /_____| \___/ \___/ |_|\_\ \___| \___| | .__/ \___| |_| (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,869] INFO | | (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,869] INFO |_| (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,869] INFO (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:zookeeper.version=3.8.2-139d619b58292d7734b4fc83a0f44be4e7b0c986, built on 2023-07-05 19:24 UTC (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:host.name=ThinkPadP53 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:java.version=17.0.6 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:java.vendor=Eclipse Adoptium (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:java.home=/opt/openjdk-bin-17.0.6_p10 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:java.class.path=/scratch/Exemple_Kafka/bin/../libs/activation-1.1.1.jar:/scratch/Exemple_Kafka/bin/../libs/aopalliance-repackaged-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/argparse4j-0.7.0.jar:/scratch/Exemple_Kafka/bin/../libs/audience-annotations-0.12.0.jar:/scratch/Exemple_Kafka/bin/../libs/caffeine-2.9.3.jar:/scratch/Exemple_Kafka/bin/../libs/checker-qual-3.19.0.jar:/scratch/Exemple_Kafka/bin/../libs/commons-beanutils-1.9.4.jar:/scratch/Exemple_Kafka/bin/../libs/commons-cli-1.4.jar:/scratch/Exemple_Kafka/bin/../libs/commons-collections-3.2.2.jar:/scratch/Exemple_Kafka/bin/../libs/commons-digester-2.1.jar:/scratch/Exemple_Kafka/bin/../libs/commons-io-2.11.0.jar:/scratch/Exemple_Kafka/bin/../libs/commons-lang3-3.8.1.jar:/scratch/Exemple_Kafka/bin/../libs/commons-logging-1.2.jar:/scratch/Exemple_Kafka/bin/../libs/commons-validator-1.7.jar:/scratch/Exemple_Kafka/bin/../libs/connect-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-basic-auth-extension-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-json-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-mirror-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-mirror-client-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-runtime-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-transforms-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/error_prone_annotations-2.10.0.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-api-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-locator-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-utils-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-annotations-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-core-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-databind-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-dataformat-csv-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-datatype-jdk8-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-jaxrs-base-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-jaxrs-json-provider-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-module-jaxb-annotations-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-module-scala_2.13-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.activation-api-1.2.2.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.annotation-api-1.3.5.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.inject-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.validation-api-2.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.ws.rs-api-2.1.6.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.xml.bind-api-2.3.3.jar:/scratch/Exemple_Kafka/bin/../libs/javassist-3.29.2-GA.jar:/scratch/Exemple_Kafka/bin/../libs/javax.activation-api-1.2.0.jar:/scratch/Exemple_Kafka/bin/../libs/javax.annotation-api-1.3.2.jar:/scratch/Exemple_Kafka/bin/../libs/javax.servlet-api-3.1.0.jar:/scratch/Exemple_Kafka/bin/../libs/javax.ws.rs-api-2.1.1.jar:/scratch/Exemple_Kafka/bin/../libs/jaxb-api-2.3.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-client-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-common-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-container-servlet-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-container-servlet-core-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-hk2-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-server-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-client-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-continuation-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-http-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-io-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-security-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-server-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-servlet-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-servlets-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-util-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-util-ajax-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jline-3.22.0.jar:/scratch/Exemple_Kafka/bin/../libs/jopt-simple-5.0.4.jar:/scratch/Exemple_Kafka/bin/../libs/jose4j-0.9.3.jar:/scratch/Exemple_Kafka/bin/../libs/jsr305-3.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-clients-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-group-coordinator-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-log4j-appender-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-metadata-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-raft-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-server-common-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-shell-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-storage-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-storage-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-examples-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-scala_2.13-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-test-utils-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-tools-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-tools-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka_2.13-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/lz4-java-1.8.0.jar:/scratch/Exemple_Kafka/bin/../libs/maven-artifact-3.8.8.jar:/scratch/Exemple_Kafka/bin/../libs/metrics-core-2.2.0.jar:/scratch/Exemple_Kafka/bin/../libs/metrics-core-4.1.12.1.jar:/scratch/Exemple_Kafka/bin/../libs/netty-buffer-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-codec-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-common-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-handler-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-resolver-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-classes-epoll-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-native-epoll-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-native-unix-common-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/osgi-resource-locator-1.0.3.jar:/scratch/Exemple_Kafka/bin/../libs/paranamer-2.8.jar:/scratch/Exemple_Kafka/bin/../libs/pcollections-4.0.1.jar:/scratch/Exemple_Kafka/bin/../libs/plexus-utils-3.3.1.jar:/scratch/Exemple_Kafka/bin/../libs/reflections-0.10.2.jar:/scratch/Exemple_Kafka/bin/../libs/reload4j-1.2.25.jar:/scratch/Exemple_Kafka/bin/../libs/rocksdbjni-7.9.2.jar:/scratch/Exemple_Kafka/bin/../libs/scala-collection-compat_2.13-2.10.0.jar:/scratch/Exemple_Kafka/bin/../libs/scala-java8-compat_2.13-1.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/scala-library-2.13.11.jar:/scratch/Exemple_Kafka/bin/../libs/scala-logging_2.13-3.9.4.jar:/scratch/Exemple_Kafka/bin/../libs/scala-reflect-2.13.11.jar:/scratch/Exemple_Kafka/bin/../libs/slf4j-api-1.7.36.jar:/scratch/Exemple_Kafka/bin/../libs/slf4j-reload4j-1.7.36.jar:/scratch/Exemple_Kafka/bin/../libs/snappy-java-1.1.10.4.jar:/scratch/Exemple_Kafka/bin/../libs/swagger-annotations-2.2.8.jar:/scratch/Exemple_Kafka/bin/../libs/trogdor-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/zookeeper-3.8.2.jar:/scratch/Exemple_Kafka/bin/../libs/zookeeper-jute-3.8.2.jar:/scratch/Exemple_Kafka/bin/../libs/zstd-jni-1.5.5-1.jar (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:java.io.tmpdir=/tmp (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:java.compiler= (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:os.name=Linux (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,872] INFO Server environment:os.arch=amd64 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO Server environment:os.version=6.4.3-cachyosGentooThinkPadP53 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO Server environment:user.name=memartel (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO Server environment:user.home=/home/memartel (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO Server environment:user.dir=/scratch/Exemple_Kafka (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO Server environment:os.memory.free=494MB (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO Server environment:os.memory.max=512MB (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO Server environment:os.memory.total=512MB (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO zookeeper.enableEagerACLCheck = false (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO zookeeper.digest.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO zookeeper.closeSessionTxn.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO zookeeper.flushDelay = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO zookeeper.maxWriteQueuePollTime = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO zookeeper.maxBatchSize=1000 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,873] INFO zookeeper.intBufferStartingSizeBytes = 1024 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,874] INFO Weighed connection throttling is disabled (org.apache.zookeeper.server.BlueThrottle) -[2023-11-03 19:38:10,875] INFO minSessionTimeout set to 6000 ms (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,876] INFO maxSessionTimeout set to 60000 ms (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,877] INFO getData response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) -[2023-11-03 19:38:10,878] INFO getChildren response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) -[2023-11-03 19:38:10,879] INFO zookeeper.pathStats.slotCapacity = 60 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:38:10,879] INFO zookeeper.pathStats.slotDuration = 15 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:38:10,879] INFO zookeeper.pathStats.maxDepth = 6 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:38:10,879] INFO zookeeper.pathStats.initialDelay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:38:10,879] INFO zookeeper.pathStats.delay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:38:10,879] INFO zookeeper.pathStats.enabled = false (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:38:10,882] INFO The max bytes for all large requests are set to 104857600 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,882] INFO The large request threshold is set to -1 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,882] INFO zookeeper.enforce.auth.enabled = false (org.apache.zookeeper.server.AuthenticationHelper) -[2023-11-03 19:38:10,882] INFO zookeeper.enforce.auth.schemes = [] (org.apache.zookeeper.server.AuthenticationHelper) -[2023-11-03 19:38:10,882] INFO Created server with tickTime 3000 ms minSessionTimeout 6000 ms maxSessionTimeout 60000 ms clientPortListenBacklog -1 datadir /tmp/zookeeper/version-2 snapdir /tmp/zookeeper/version-2 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,887] INFO Using org.apache.zookeeper.server.NIOServerCnxnFactory as server connection factory (org.apache.zookeeper.server.ServerCnxnFactory) -[2023-11-03 19:38:10,888] WARN maxCnxns is not configured, using default value 0. (org.apache.zookeeper.server.ServerCnxnFactory) -[2023-11-03 19:38:10,889] INFO Configuring NIO connection handler with 10s sessionless connection timeout, 2 selector thread(s), 24 worker threads, and 64 kB direct buffers. (org.apache.zookeeper.server.NIOServerCnxnFactory) -[2023-11-03 19:38:10,896] INFO binding to port 0.0.0.0/0.0.0.0:2181 (org.apache.zookeeper.server.NIOServerCnxnFactory) -[2023-11-03 19:38:10,908] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) -[2023-11-03 19:38:10,908] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) -[2023-11-03 19:38:10,908] INFO zookeeper.snapshotSizeFactor = 0.33 (org.apache.zookeeper.server.ZKDatabase) -[2023-11-03 19:38:10,908] INFO zookeeper.commitLogCount=500 (org.apache.zookeeper.server.ZKDatabase) -[2023-11-03 19:38:10,914] INFO zookeeper.snapshot.compression.method = CHECKED (org.apache.zookeeper.server.persistence.SnapStream) -[2023-11-03 19:38:10,914] INFO Snapshotting: 0x0 to /tmp/zookeeper/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog) -[2023-11-03 19:38:10,917] INFO Snapshot loaded in 8 ms, highest zxid is 0x0, digest is 1371985504 (org.apache.zookeeper.server.ZKDatabase) -[2023-11-03 19:38:10,918] INFO Snapshotting: 0x0 to /tmp/zookeeper/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog) -[2023-11-03 19:38:10,918] INFO Snapshot taken in 1 ms (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:38:10,926] INFO PrepRequestProcessor (sid:0) started, reconfigEnabled=false (org.apache.zookeeper.server.PrepRequestProcessor) -[2023-11-03 19:38:10,926] INFO zookeeper.request_throttler.shutdownTimeout = 10000 ms (org.apache.zookeeper.server.RequestThrottler) -[2023-11-03 19:38:10,939] INFO Using checkIntervalMs=60000 maxPerMinute=10000 maxNeverUsedIntervalMs=0 (org.apache.zookeeper.server.ContainerManager) -[2023-11-03 19:38:10,940] INFO ZooKeeper audit is disabled. (org.apache.zookeeper.audit.ZKAuditProvider) -[2023-11-03 19:38:20,073] INFO Registered kafka:type=kafka.Log4jController MBean (kafka.utils.Log4jControllerRegistration$) -[2023-11-03 19:38:20,292] INFO Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation (org.apache.zookeeper.common.X509Util) -[2023-11-03 19:38:20,354] INFO Registered signal handlers for TERM, INT, HUP (org.apache.kafka.common.utils.LoggingSignalHandler) -[2023-11-03 19:38:20,355] INFO starting (kafka.server.KafkaServer) -[2023-11-03 19:38:20,355] INFO Connecting to zookeeper on localhost:2181 (kafka.server.KafkaServer) -[2023-11-03 19:38:20,366] INFO [ZooKeeperClient Kafka server] Initializing a new session to localhost:2181. (kafka.zookeeper.ZooKeeperClient) -[2023-11-03 19:38:20,371] INFO Client environment:zookeeper.version=3.8.2-139d619b58292d7734b4fc83a0f44be4e7b0c986, built on 2023-07-05 19:24 UTC (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:host.name=ThinkPadP53 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:java.version=17.0.6 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:java.vendor=Eclipse Adoptium (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:java.home=/opt/openjdk-bin-17.0.6_p10 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:java.class.path=/scratch/Exemple_Kafka/bin/../libs/activation-1.1.1.jar:/scratch/Exemple_Kafka/bin/../libs/aopalliance-repackaged-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/argparse4j-0.7.0.jar:/scratch/Exemple_Kafka/bin/../libs/audience-annotations-0.12.0.jar:/scratch/Exemple_Kafka/bin/../libs/caffeine-2.9.3.jar:/scratch/Exemple_Kafka/bin/../libs/checker-qual-3.19.0.jar:/scratch/Exemple_Kafka/bin/../libs/commons-beanutils-1.9.4.jar:/scratch/Exemple_Kafka/bin/../libs/commons-cli-1.4.jar:/scratch/Exemple_Kafka/bin/../libs/commons-collections-3.2.2.jar:/scratch/Exemple_Kafka/bin/../libs/commons-digester-2.1.jar:/scratch/Exemple_Kafka/bin/../libs/commons-io-2.11.0.jar:/scratch/Exemple_Kafka/bin/../libs/commons-lang3-3.8.1.jar:/scratch/Exemple_Kafka/bin/../libs/commons-logging-1.2.jar:/scratch/Exemple_Kafka/bin/../libs/commons-validator-1.7.jar:/scratch/Exemple_Kafka/bin/../libs/connect-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-basic-auth-extension-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-json-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-mirror-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-mirror-client-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-runtime-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-transforms-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/error_prone_annotations-2.10.0.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-api-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-locator-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-utils-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-annotations-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-core-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-databind-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-dataformat-csv-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-datatype-jdk8-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-jaxrs-base-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-jaxrs-json-provider-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-module-jaxb-annotations-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-module-scala_2.13-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.activation-api-1.2.2.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.annotation-api-1.3.5.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.inject-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.validation-api-2.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.ws.rs-api-2.1.6.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.xml.bind-api-2.3.3.jar:/scratch/Exemple_Kafka/bin/../libs/javassist-3.29.2-GA.jar:/scratch/Exemple_Kafka/bin/../libs/javax.activation-api-1.2.0.jar:/scratch/Exemple_Kafka/bin/../libs/javax.annotation-api-1.3.2.jar:/scratch/Exemple_Kafka/bin/../libs/javax.servlet-api-3.1.0.jar:/scratch/Exemple_Kafka/bin/../libs/javax.ws.rs-api-2.1.1.jar:/scratch/Exemple_Kafka/bin/../libs/jaxb-api-2.3.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-client-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-common-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-container-servlet-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-container-servlet-core-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-hk2-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-server-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-client-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-continuation-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-http-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-io-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-security-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-server-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-servlet-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-servlets-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-util-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-util-ajax-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jline-3.22.0.jar:/scratch/Exemple_Kafka/bin/../libs/jopt-simple-5.0.4.jar:/scratch/Exemple_Kafka/bin/../libs/jose4j-0.9.3.jar:/scratch/Exemple_Kafka/bin/../libs/jsr305-3.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-clients-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-group-coordinator-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-log4j-appender-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-metadata-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-raft-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-server-common-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-shell-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-storage-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-storage-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-examples-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-scala_2.13-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-test-utils-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-tools-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-tools-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka_2.13-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/lz4-java-1.8.0.jar:/scratch/Exemple_Kafka/bin/../libs/maven-artifact-3.8.8.jar:/scratch/Exemple_Kafka/bin/../libs/metrics-core-2.2.0.jar:/scratch/Exemple_Kafka/bin/../libs/metrics-core-4.1.12.1.jar:/scratch/Exemple_Kafka/bin/../libs/netty-buffer-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-codec-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-common-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-handler-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-resolver-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-classes-epoll-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-native-epoll-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-native-unix-common-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/osgi-resource-locator-1.0.3.jar:/scratch/Exemple_Kafka/bin/../libs/paranamer-2.8.jar:/scratch/Exemple_Kafka/bin/../libs/pcollections-4.0.1.jar:/scratch/Exemple_Kafka/bin/../libs/plexus-utils-3.3.1.jar:/scratch/Exemple_Kafka/bin/../libs/reflections-0.10.2.jar:/scratch/Exemple_Kafka/bin/../libs/reload4j-1.2.25.jar:/scratch/Exemple_Kafka/bin/../libs/rocksdbjni-7.9.2.jar:/scratch/Exemple_Kafka/bin/../libs/scala-collection-compat_2.13-2.10.0.jar:/scratch/Exemple_Kafka/bin/../libs/scala-java8-compat_2.13-1.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/scala-library-2.13.11.jar:/scratch/Exemple_Kafka/bin/../libs/scala-logging_2.13-3.9.4.jar:/scratch/Exemple_Kafka/bin/../libs/scala-reflect-2.13.11.jar:/scratch/Exemple_Kafka/bin/../libs/slf4j-api-1.7.36.jar:/scratch/Exemple_Kafka/bin/../libs/slf4j-reload4j-1.7.36.jar:/scratch/Exemple_Kafka/bin/../libs/snappy-java-1.1.10.4.jar:/scratch/Exemple_Kafka/bin/../libs/swagger-annotations-2.2.8.jar:/scratch/Exemple_Kafka/bin/../libs/trogdor-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/zookeeper-3.8.2.jar:/scratch/Exemple_Kafka/bin/../libs/zookeeper-jute-3.8.2.jar:/scratch/Exemple_Kafka/bin/../libs/zstd-jni-1.5.5-1.jar (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:java.compiler= (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:os.version=6.4.3-cachyosGentooThinkPadP53 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:user.name=memartel (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,371] INFO Client environment:user.home=/home/memartel (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,372] INFO Client environment:user.dir=/scratch/Exemple_Kafka (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,372] INFO Client environment:os.memory.free=987MB (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,372] INFO Client environment:os.memory.max=1024MB (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,372] INFO Client environment:os.memory.total=1024MB (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,374] INFO Initiating client connection, connectString=localhost:2181 sessionTimeout=18000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@3fce8fd9 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:38:20,381] INFO jute.maxbuffer value is 4194304 Bytes (org.apache.zookeeper.ClientCnxnSocket) -[2023-11-03 19:38:20,385] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn) -[2023-11-03 19:38:20,386] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient) -[2023-11-03 19:38:20,387] INFO Opening socket connection to server localhost/127.0.0.1:2181. (org.apache.zookeeper.ClientCnxn) -[2023-11-03 19:38:20,389] INFO Socket connection established, initiating session, client: /127.0.0.1:41842, server: localhost/127.0.0.1:2181 (org.apache.zookeeper.ClientCnxn) -[2023-11-03 19:38:20,397] INFO Creating new log file: log.1 (org.apache.zookeeper.server.persistence.FileTxnLog) -[2023-11-03 19:38:20,402] INFO Session establishment complete on server localhost/127.0.0.1:2181, session id = 0x100008497380000, negotiated timeout = 18000 (org.apache.zookeeper.ClientCnxn) -[2023-11-03 19:38:20,404] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient) -[2023-11-03 19:38:20,606] INFO Cluster ID = 8EdUX-SkTwmB2aSLdVSXoQ (kafka.server.KafkaServer) -[2023-11-03 19:38:20,608] WARN No meta.properties file under dir /tmp/kafka-logs/meta.properties (kafka.server.BrokerMetadataCheckpoint) -[2023-11-03 19:38:20,639] INFO KafkaConfig values: +[2023-11-06 13:26:24,196] INFO Reading configuration from: ./config/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) +[2023-11-06 13:26:24,198] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) +[2023-11-06 13:26:24,198] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) +[2023-11-06 13:26:24,198] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) +[2023-11-06 13:26:24,198] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) +[2023-11-06 13:26:24,199] INFO autopurge.snapRetainCount set to 3 (org.apache.zookeeper.server.DatadirCleanupManager) +[2023-11-06 13:26:24,199] INFO autopurge.purgeInterval set to 0 (org.apache.zookeeper.server.DatadirCleanupManager) +[2023-11-06 13:26:24,200] INFO Purge task is not scheduled. (org.apache.zookeeper.server.DatadirCleanupManager) +[2023-11-06 13:26:24,200] WARN Either no config or no quorum defined in config, running in standalone mode (org.apache.zookeeper.server.quorum.QuorumPeerMain) +[2023-11-06 13:26:24,201] INFO Log4j 1.2 jmx support not found; jmx disabled. (org.apache.zookeeper.jmx.ManagedUtil) +[2023-11-06 13:26:24,201] INFO Reading configuration from: ./config/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) +[2023-11-06 13:26:24,202] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) +[2023-11-06 13:26:24,202] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) +[2023-11-06 13:26:24,202] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) +[2023-11-06 13:26:24,202] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) +[2023-11-06 13:26:24,202] INFO Starting server (org.apache.zookeeper.server.ZooKeeperServerMain) +[2023-11-06 13:26:24,212] INFO ServerMetrics initialized with provider org.apache.zookeeper.metrics.impl.DefaultMetricsProvider@e50a6f6 (org.apache.zookeeper.server.ServerMetrics) +[2023-11-06 13:26:24,214] INFO ACL digest algorithm is: SHA1 (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) +[2023-11-06 13:26:24,214] INFO zookeeper.DigestAuthenticationProvider.enabled = true (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) +[2023-11-06 13:26:24,217] INFO zookeeper.snapshot.trust.empty : false (org.apache.zookeeper.server.persistence.FileTxnSnapLog) +[2023-11-06 13:26:24,224] INFO (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,224] INFO ______ _ (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,224] INFO |___ / | | (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,224] INFO / / ___ ___ | | __ ___ ___ _ __ ___ _ __ (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,224] INFO / / / _ \ / _ \ | |/ / / _ \ / _ \ | '_ \ / _ \ | '__| (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,224] INFO / /__ | (_) | | (_) | | < | __/ | __/ | |_) | | __/ | | (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,224] INFO /_____| \___/ \___/ |_|\_\ \___| \___| | .__/ \___| |_| (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,224] INFO | | (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,224] INFO |_| (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,224] INFO (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,225] INFO Server environment:zookeeper.version=3.8.2-139d619b58292d7734b4fc83a0f44be4e7b0c986, built on 2023-07-05 19:24 UTC (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,225] INFO Server environment:host.name=localhost (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,225] INFO Server environment:java.version=17.0.8.1 (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,225] INFO Server environment:java.vendor=Eclipse Adoptium (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,226] INFO Server environment:java.home=/opt/openjdk-bin-17.0.8.1_p1 (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,226] INFO Server environment:java.class.path=/scratch/repos/Exemple_Kafka/bin/../libs/activation-1.1.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/aopalliance-repackaged-2.6.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/argparse4j-0.7.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/audience-annotations-0.12.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/caffeine-2.9.3.jar:/scratch/repos/Exemple_Kafka/bin/../libs/checker-qual-3.19.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-beanutils-1.9.4.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-cli-1.4.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-collections-3.2.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-digester-2.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-io-2.11.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-lang3-3.8.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-logging-1.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-validator-1.7.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-api-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-basic-auth-extension-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-json-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-mirror-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-mirror-client-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-runtime-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-transforms-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/error_prone_annotations-2.10.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/hk2-api-2.6.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/hk2-locator-2.6.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/hk2-utils-2.6.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-annotations-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-core-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-databind-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-dataformat-csv-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-datatype-jdk8-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-jaxrs-base-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-jaxrs-json-provider-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-module-jaxb-annotations-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-module-scala_2.13-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.activation-api-1.2.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.annotation-api-1.3.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.inject-2.6.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.validation-api-2.0.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.ws.rs-api-2.1.6.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.xml.bind-api-2.3.3.jar:/scratch/repos/Exemple_Kafka/bin/../libs/javassist-3.29.2-GA.jar:/scratch/repos/Exemple_Kafka/bin/../libs/javax.activation-api-1.2.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/javax.annotation-api-1.3.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/javax.servlet-api-3.1.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/javax.ws.rs-api-2.1.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jaxb-api-2.3.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-client-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-common-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-container-servlet-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-container-servlet-core-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-hk2-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-server-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-client-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-continuation-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-http-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-io-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-security-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-server-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-servlet-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-servlets-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-util-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-util-ajax-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jline-3.22.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jopt-simple-5.0.4.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jose4j-0.9.3.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jsr305-3.0.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-clients-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-group-coordinator-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-log4j-appender-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-metadata-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-raft-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-server-common-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-shell-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-storage-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-storage-api-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-streams-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-streams-examples-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-streams-scala_2.13-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-streams-test-utils-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-tools-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-tools-api-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka_2.13-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/lz4-java-1.8.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/maven-artifact-3.8.8.jar:/scratch/repos/Exemple_Kafka/bin/../libs/metrics-core-2.2.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/metrics-core-4.1.12.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-buffer-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-codec-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-common-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-handler-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-resolver-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-transport-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-transport-classes-epoll-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-transport-native-epoll-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-transport-native-unix-common-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/osgi-resource-locator-1.0.3.jar:/scratch/repos/Exemple_Kafka/bin/../libs/paranamer-2.8.jar:/scratch/repos/Exemple_Kafka/bin/../libs/pcollections-4.0.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/plexus-utils-3.3.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/reflections-0.10.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/reload4j-1.2.25.jar:/scratch/repos/Exemple_Kafka/bin/../libs/rocksdbjni-7.9.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/scala-collection-compat_2.13-2.10.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/scala-java8-compat_2.13-1.0.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/scala-library-2.13.11.jar:/scratch/repos/Exemple_Kafka/bin/../libs/scala-logging_2.13-3.9.4.jar:/scratch/repos/Exemple_Kafka/bin/../libs/scala-reflect-2.13.11.jar:/scratch/repos/Exemple_Kafka/bin/../libs/slf4j-api-1.7.36.jar:/scratch/repos/Exemple_Kafka/bin/../libs/slf4j-reload4j-1.7.36.jar:/scratch/repos/Exemple_Kafka/bin/../libs/snappy-java-1.1.10.4.jar:/scratch/repos/Exemple_Kafka/bin/../libs/swagger-annotations-2.2.8.jar:/scratch/repos/Exemple_Kafka/bin/../libs/trogdor-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/zookeeper-3.8.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/zookeeper-jute-3.8.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/zstd-jni-1.5.5-1.jar (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,226] INFO Server environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,226] INFO Server environment:java.io.tmpdir=/tmp (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,226] INFO Server environment:java.compiler= (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,226] INFO Server environment:os.name=Linux (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,226] INFO Server environment:os.arch=amd64 (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,226] INFO Server environment:os.version=6.4.3-cachyosGentooThinkPadP53 (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,226] INFO Server environment:user.name=memartel (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO Server environment:user.home=/home/memartel (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO Server environment:user.dir=/scratch/repos/Exemple_Kafka (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO Server environment:os.memory.free=494MB (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO Server environment:os.memory.max=512MB (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO Server environment:os.memory.total=512MB (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO zookeeper.enableEagerACLCheck = false (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO zookeeper.digest.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO zookeeper.closeSessionTxn.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO zookeeper.flushDelay = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO zookeeper.maxWriteQueuePollTime = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO zookeeper.maxBatchSize=1000 (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,227] INFO zookeeper.intBufferStartingSizeBytes = 1024 (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,228] INFO Weighed connection throttling is disabled (org.apache.zookeeper.server.BlueThrottle) +[2023-11-06 13:26:24,229] INFO minSessionTimeout set to 6000 ms (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,229] INFO maxSessionTimeout set to 60000 ms (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,230] INFO getData response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) +[2023-11-06 13:26:24,230] INFO getChildren response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) +[2023-11-06 13:26:24,231] INFO zookeeper.pathStats.slotCapacity = 60 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) +[2023-11-06 13:26:24,231] INFO zookeeper.pathStats.slotDuration = 15 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) +[2023-11-06 13:26:24,231] INFO zookeeper.pathStats.maxDepth = 6 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) +[2023-11-06 13:26:24,231] INFO zookeeper.pathStats.initialDelay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) +[2023-11-06 13:26:24,231] INFO zookeeper.pathStats.delay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) +[2023-11-06 13:26:24,231] INFO zookeeper.pathStats.enabled = false (org.apache.zookeeper.server.util.RequestPathMetricsCollector) +[2023-11-06 13:26:24,233] INFO The max bytes for all large requests are set to 104857600 (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,233] INFO The large request threshold is set to -1 (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,234] INFO zookeeper.enforce.auth.enabled = false (org.apache.zookeeper.server.AuthenticationHelper) +[2023-11-06 13:26:24,234] INFO zookeeper.enforce.auth.schemes = [] (org.apache.zookeeper.server.AuthenticationHelper) +[2023-11-06 13:26:24,234] INFO Created server with tickTime 3000 ms minSessionTimeout 6000 ms maxSessionTimeout 60000 ms clientPortListenBacklog -1 datadir /tmp/zookeeper/version-2 snapdir /tmp/zookeeper/version-2 (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,238] INFO Using org.apache.zookeeper.server.NIOServerCnxnFactory as server connection factory (org.apache.zookeeper.server.ServerCnxnFactory) +[2023-11-06 13:26:24,238] WARN maxCnxns is not configured, using default value 0. (org.apache.zookeeper.server.ServerCnxnFactory) +[2023-11-06 13:26:24,239] INFO Configuring NIO connection handler with 10s sessionless connection timeout, 2 selector thread(s), 24 worker threads, and 64 kB direct buffers. (org.apache.zookeeper.server.NIOServerCnxnFactory) +[2023-11-06 13:26:24,244] INFO binding to port 0.0.0.0/0.0.0.0:2181 (org.apache.zookeeper.server.NIOServerCnxnFactory) +[2023-11-06 13:26:24,254] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) +[2023-11-06 13:26:24,254] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) +[2023-11-06 13:26:24,254] INFO zookeeper.snapshotSizeFactor = 0.33 (org.apache.zookeeper.server.ZKDatabase) +[2023-11-06 13:26:24,254] INFO zookeeper.commitLogCount=500 (org.apache.zookeeper.server.ZKDatabase) +[2023-11-06 13:26:24,260] INFO zookeeper.snapshot.compression.method = CHECKED (org.apache.zookeeper.server.persistence.SnapStream) +[2023-11-06 13:26:24,260] INFO Snapshotting: 0x0 to /tmp/zookeeper/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog) +[2023-11-06 13:26:24,262] INFO Snapshot loaded in 8 ms, highest zxid is 0x0, digest is 1371985504 (org.apache.zookeeper.server.ZKDatabase) +[2023-11-06 13:26:24,263] INFO Snapshotting: 0x0 to /tmp/zookeeper/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileTxnSnapLog) +[2023-11-06 13:26:24,263] INFO Snapshot taken in 0 ms (org.apache.zookeeper.server.ZooKeeperServer) +[2023-11-06 13:26:24,270] INFO PrepRequestProcessor (sid:0) started, reconfigEnabled=false (org.apache.zookeeper.server.PrepRequestProcessor) +[2023-11-06 13:26:24,271] INFO zookeeper.request_throttler.shutdownTimeout = 10000 ms (org.apache.zookeeper.server.RequestThrottler) +[2023-11-06 13:26:24,282] INFO Using checkIntervalMs=60000 maxPerMinute=10000 maxNeverUsedIntervalMs=0 (org.apache.zookeeper.server.ContainerManager) +[2023-11-06 13:26:24,282] INFO ZooKeeper audit is disabled. (org.apache.zookeeper.audit.ZKAuditProvider) +[2023-11-06 13:26:31,049] INFO Registered kafka:type=kafka.Log4jController MBean (kafka.utils.Log4jControllerRegistration$) +[2023-11-06 13:26:31,233] INFO Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation (org.apache.zookeeper.common.X509Util) +[2023-11-06 13:26:31,292] INFO Registered signal handlers for TERM, INT, HUP (org.apache.kafka.common.utils.LoggingSignalHandler) +[2023-11-06 13:26:31,293] INFO starting (kafka.server.KafkaServer) +[2023-11-06 13:26:31,293] INFO Connecting to zookeeper on localhost:2181 (kafka.server.KafkaServer) +[2023-11-06 13:26:31,304] INFO [ZooKeeperClient Kafka server] Initializing a new session to localhost:2181. (kafka.zookeeper.ZooKeeperClient) +[2023-11-06 13:26:31,308] INFO Client environment:zookeeper.version=3.8.2-139d619b58292d7734b4fc83a0f44be4e7b0c986, built on 2023-07-05 19:24 UTC (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,308] INFO Client environment:host.name=localhost (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,308] INFO Client environment:java.version=17.0.8.1 (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,308] INFO Client environment:java.vendor=Eclipse Adoptium (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,308] INFO Client environment:java.home=/opt/openjdk-bin-17.0.8.1_p1 (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,308] INFO Client environment:java.class.path=/scratch/repos/Exemple_Kafka/bin/../libs/activation-1.1.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/aopalliance-repackaged-2.6.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/argparse4j-0.7.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/audience-annotations-0.12.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/caffeine-2.9.3.jar:/scratch/repos/Exemple_Kafka/bin/../libs/checker-qual-3.19.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-beanutils-1.9.4.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-cli-1.4.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-collections-3.2.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-digester-2.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-io-2.11.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-lang3-3.8.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-logging-1.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/commons-validator-1.7.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-api-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-basic-auth-extension-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-json-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-mirror-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-mirror-client-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-runtime-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/connect-transforms-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/error_prone_annotations-2.10.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/hk2-api-2.6.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/hk2-locator-2.6.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/hk2-utils-2.6.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-annotations-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-core-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-databind-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-dataformat-csv-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-datatype-jdk8-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-jaxrs-base-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-jaxrs-json-provider-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-module-jaxb-annotations-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jackson-module-scala_2.13-2.13.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.activation-api-1.2.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.annotation-api-1.3.5.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.inject-2.6.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.validation-api-2.0.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.ws.rs-api-2.1.6.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jakarta.xml.bind-api-2.3.3.jar:/scratch/repos/Exemple_Kafka/bin/../libs/javassist-3.29.2-GA.jar:/scratch/repos/Exemple_Kafka/bin/../libs/javax.activation-api-1.2.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/javax.annotation-api-1.3.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/javax.servlet-api-3.1.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/javax.ws.rs-api-2.1.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jaxb-api-2.3.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-client-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-common-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-container-servlet-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-container-servlet-core-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-hk2-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jersey-server-2.39.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-client-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-continuation-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-http-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-io-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-security-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-server-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-servlet-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-servlets-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-util-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jetty-util-ajax-9.4.52.v20230823.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jline-3.22.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jopt-simple-5.0.4.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jose4j-0.9.3.jar:/scratch/repos/Exemple_Kafka/bin/../libs/jsr305-3.0.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-clients-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-group-coordinator-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-log4j-appender-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-metadata-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-raft-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-server-common-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-shell-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-storage-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-storage-api-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-streams-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-streams-examples-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-streams-scala_2.13-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-streams-test-utils-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-tools-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka-tools-api-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/kafka_2.13-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/lz4-java-1.8.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/maven-artifact-3.8.8.jar:/scratch/repos/Exemple_Kafka/bin/../libs/metrics-core-2.2.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/metrics-core-4.1.12.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-buffer-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-codec-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-common-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-handler-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-resolver-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-transport-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-transport-classes-epoll-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-transport-native-epoll-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/netty-transport-native-unix-common-4.1.94.Final.jar:/scratch/repos/Exemple_Kafka/bin/../libs/osgi-resource-locator-1.0.3.jar:/scratch/repos/Exemple_Kafka/bin/../libs/paranamer-2.8.jar:/scratch/repos/Exemple_Kafka/bin/../libs/pcollections-4.0.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/plexus-utils-3.3.1.jar:/scratch/repos/Exemple_Kafka/bin/../libs/reflections-0.10.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/reload4j-1.2.25.jar:/scratch/repos/Exemple_Kafka/bin/../libs/rocksdbjni-7.9.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/scala-collection-compat_2.13-2.10.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/scala-java8-compat_2.13-1.0.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/scala-library-2.13.11.jar:/scratch/repos/Exemple_Kafka/bin/../libs/scala-logging_2.13-3.9.4.jar:/scratch/repos/Exemple_Kafka/bin/../libs/scala-reflect-2.13.11.jar:/scratch/repos/Exemple_Kafka/bin/../libs/slf4j-api-1.7.36.jar:/scratch/repos/Exemple_Kafka/bin/../libs/slf4j-reload4j-1.7.36.jar:/scratch/repos/Exemple_Kafka/bin/../libs/snappy-java-1.1.10.4.jar:/scratch/repos/Exemple_Kafka/bin/../libs/swagger-annotations-2.2.8.jar:/scratch/repos/Exemple_Kafka/bin/../libs/trogdor-3.6.0.jar:/scratch/repos/Exemple_Kafka/bin/../libs/zookeeper-3.8.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/zookeeper-jute-3.8.2.jar:/scratch/repos/Exemple_Kafka/bin/../libs/zstd-jni-1.5.5-1.jar (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:java.compiler= (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:os.version=6.4.3-cachyosGentooThinkPadP53 (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:user.name=memartel (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:user.home=/home/memartel (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:user.dir=/scratch/repos/Exemple_Kafka (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:os.memory.free=987MB (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:os.memory.max=1024MB (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,309] INFO Client environment:os.memory.total=1024MB (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,311] INFO Initiating client connection, connectString=localhost:2181 sessionTimeout=18000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@3fce8fd9 (org.apache.zookeeper.ZooKeeper) +[2023-11-06 13:26:31,316] INFO jute.maxbuffer value is 4194304 Bytes (org.apache.zookeeper.ClientCnxnSocket) +[2023-11-06 13:26:31,321] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn) +[2023-11-06 13:26:31,322] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient) +[2023-11-06 13:26:31,323] INFO Opening socket connection to server localhost/[0:0:0:0:0:0:0:1]:2181. (org.apache.zookeeper.ClientCnxn) +[2023-11-06 13:26:31,325] INFO Socket connection established, initiating session, client: /[0:0:0:0:0:0:0:1]:60976, server: localhost/[0:0:0:0:0:0:0:1]:2181 (org.apache.zookeeper.ClientCnxn) +[2023-11-06 13:26:31,331] INFO Creating new log file: log.1 (org.apache.zookeeper.server.persistence.FileTxnLog) +[2023-11-06 13:26:31,336] INFO Session establishment complete on server localhost/[0:0:0:0:0:0:0:1]:2181, session id = 0x1000000d8b20000, negotiated timeout = 18000 (org.apache.zookeeper.ClientCnxn) +[2023-11-06 13:26:31,338] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient) +[2023-11-06 13:26:31,519] INFO Cluster ID = tHet8_ZjRl6pr3Pp3utDPw (kafka.server.KafkaServer) +[2023-11-06 13:26:31,521] WARN No meta.properties file under dir /tmp/kafka-logs/meta.properties (kafka.server.BrokerMetadataCheckpoint) +[2023-11-06 13:26:31,553] INFO KafkaConfig values: advertised.listeners = null alter.config.policy.class.name = null alter.log.dirs.replication.quota.window.num = 11 @@ -411,1299 +411,417 @@ zookeeper.ssl.truststore.password = null zookeeper.ssl.truststore.type = null (kafka.server.KafkaConfig) -[2023-11-03 19:38:20,668] INFO [ThrottledChannelReaper-Fetch]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:38:20,668] INFO [ThrottledChannelReaper-Produce]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:38:20,669] INFO [ThrottledChannelReaper-Request]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:38:20,670] INFO [ThrottledChannelReaper-ControllerMutation]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:38:20,683] INFO Log directory /tmp/kafka-logs not found, creating it. (kafka.log.LogManager) -[2023-11-03 19:38:20,695] INFO Loading logs from log dirs ArraySeq(/tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:38:20,698] INFO No logs found to be loaded in /tmp/kafka-logs (kafka.log.LogManager) -[2023-11-03 19:38:20,704] INFO Loaded 0 logs in 9ms (kafka.log.LogManager) -[2023-11-03 19:38:20,705] INFO Starting log cleanup with a period of 300000 ms. (kafka.log.LogManager) -[2023-11-03 19:38:20,706] INFO Starting log flusher with a default period of 9223372036854775807 ms. (kafka.log.LogManager) -[2023-11-03 19:38:20,739] INFO [kafka-log-cleaner-thread-0]: Starting (kafka.log.LogCleaner$CleanerThread) -[2023-11-03 19:38:20,749] INFO [feature-zk-node-event-process-thread]: Starting (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) -[2023-11-03 19:38:20,755] INFO Feature ZK node at path: /feature does not exist (kafka.server.FinalizedFeatureChangeListener) -[2023-11-03 19:38:20,788] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Starting (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:38:20,988] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) -[2023-11-03 19:38:21,001] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer) -[2023-11-03 19:38:21,004] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Starting (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:38:21,021] INFO [ExpirationReaper-0-Produce]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:38:21,022] INFO [ExpirationReaper-0-Fetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:38:21,023] INFO [ExpirationReaper-0-DeleteRecords]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:38:21,024] INFO [ExpirationReaper-0-ElectLeader]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:38:21,025] INFO [ExpirationReaper-0-RemoteFetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:38:21,036] INFO [LogDirFailureHandler]: Starting (kafka.server.ReplicaManager$LogDirFailureHandler) -[2023-11-03 19:38:21,037] INFO [AddPartitionsToTxnSenderThread-0]: Starting (kafka.server.AddPartitionsToTxnManager) -[2023-11-03 19:38:21,064] INFO Creating /brokers/ids/0 (is it secure? false) (kafka.zk.KafkaZkClient) -[2023-11-03 19:38:21,083] INFO Stat of the created znode at /brokers/ids/0 is: 25,25,1699054701074,1699054701074,1,0,0,72058163510640640,206,0,25 +[2023-11-06 13:26:31,577] INFO [ThrottledChannelReaper-Fetch]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) +[2023-11-06 13:26:31,577] INFO [ThrottledChannelReaper-Produce]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) +[2023-11-06 13:26:31,578] INFO [ThrottledChannelReaper-Request]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) +[2023-11-06 13:26:31,580] INFO [ThrottledChannelReaper-ControllerMutation]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) +[2023-11-06 13:26:31,588] INFO Log directory /tmp/kafka-logs not found, creating it. (kafka.log.LogManager) +[2023-11-06 13:26:31,599] INFO Loading logs from log dirs ArraySeq(/tmp/kafka-logs) (kafka.log.LogManager) +[2023-11-06 13:26:31,601] INFO No logs found to be loaded in /tmp/kafka-logs (kafka.log.LogManager) +[2023-11-06 13:26:31,607] INFO Loaded 0 logs in 8ms (kafka.log.LogManager) +[2023-11-06 13:26:31,608] INFO Starting log cleanup with a period of 300000 ms. (kafka.log.LogManager) +[2023-11-06 13:26:31,609] INFO Starting log flusher with a default period of 9223372036854775807 ms. (kafka.log.LogManager) +[2023-11-06 13:26:31,634] INFO [kafka-log-cleaner-thread-0]: Starting (kafka.log.LogCleaner$CleanerThread) +[2023-11-06 13:26:31,642] INFO [feature-zk-node-event-process-thread]: Starting (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) +[2023-11-06 13:26:31,649] INFO Feature ZK node at path: /feature does not exist (kafka.server.FinalizedFeatureChangeListener) +[2023-11-06 13:26:31,674] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Starting (kafka.server.BrokerToControllerRequestThread) +[2023-11-06 13:26:31,857] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) +[2023-11-06 13:26:31,868] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer) +[2023-11-06 13:26:31,871] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Starting (kafka.server.BrokerToControllerRequestThread) +[2023-11-06 13:26:31,885] INFO [ExpirationReaper-0-Produce]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) +[2023-11-06 13:26:31,886] INFO [ExpirationReaper-0-Fetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) +[2023-11-06 13:26:31,888] INFO [ExpirationReaper-0-DeleteRecords]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) +[2023-11-06 13:26:31,888] INFO [ExpirationReaper-0-ElectLeader]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) +[2023-11-06 13:26:31,889] INFO [ExpirationReaper-0-RemoteFetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) +[2023-11-06 13:26:31,897] INFO [LogDirFailureHandler]: Starting (kafka.server.ReplicaManager$LogDirFailureHandler) +[2023-11-06 13:26:31,898] INFO [AddPartitionsToTxnSenderThread-0]: Starting (kafka.server.AddPartitionsToTxnManager) +[2023-11-06 13:26:31,911] INFO Creating /brokers/ids/0 (is it secure? false) (kafka.zk.KafkaZkClient) +[2023-11-06 13:26:31,924] INFO Stat of the created znode at /brokers/ids/0 is: 25,25,1699295191920,1699295191920,1,0,0,72057597673472000,202,0,25 (kafka.zk.KafkaZkClient) -[2023-11-03 19:38:21,083] INFO Registered broker 0 at path /brokers/ids/0 with addresses: PLAINTEXT://ThinkPadP53:9092, czxid (broker epoch): 25 (kafka.zk.KafkaZkClient) -[2023-11-03 19:38:21,127] INFO [ExpirationReaper-0-topic]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:38:21,133] INFO [ExpirationReaper-0-Heartbeat]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:38:21,133] INFO [ExpirationReaper-0-Rebalance]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:38:21,134] INFO Successfully created /controller_epoch with initial epoch 0 (kafka.zk.KafkaZkClient) -[2023-11-03 19:38:21,147] INFO [GroupCoordinator 0]: Starting up. (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:21,148] INFO Feature ZK node created at path: /feature (kafka.server.FinalizedFeatureChangeListener) -[2023-11-03 19:38:21,152] INFO [GroupCoordinator 0]: Startup complete. (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:21,164] INFO [TransactionCoordinator id=0] Starting up. (kafka.coordinator.transaction.TransactionCoordinator) -[2023-11-03 19:38:21,167] INFO [MetadataCache brokerId=0] Updated cache from existing None to latest Features(version=3.6-IV2, finalizedFeatures={}, finalizedFeaturesEpoch=0). (kafka.server.metadata.ZkMetadataCache) -[2023-11-03 19:38:21,167] INFO [TxnMarkerSenderThread-0]: Starting (kafka.coordinator.transaction.TransactionMarkerChannelManager) -[2023-11-03 19:38:21,168] INFO [TransactionCoordinator id=0] Startup complete. (kafka.coordinator.transaction.TransactionCoordinator) -[2023-11-03 19:38:21,199] INFO [ExpirationReaper-0-AlterAcls]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:38:21,219] INFO [/config/changes-event-process-thread]: Starting (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread) -[2023-11-03 19:38:21,232] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Enabling request processing. (kafka.network.SocketServer) -[2023-11-03 19:38:21,243] INFO [Controller id=0, targetBrokerId=0] Node 0 disconnected. (org.apache.kafka.clients.NetworkClient) -[2023-11-03 19:38:21,243] INFO Awaiting socket connections on 0.0.0.0:9092. (kafka.network.DataPlaneAcceptor) -[2023-11-03 19:38:21,245] WARN [Controller id=0, targetBrokerId=0] Connection to node 0 (ThinkPadP53/192.168.37.3:9092) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient) -[2023-11-03 19:38:21,247] INFO [Controller id=0, targetBrokerId=0] Client requested connection close from node 0 (org.apache.kafka.clients.NetworkClient) -[2023-11-03 19:38:21,249] INFO Kafka version: 3.6.0 (org.apache.kafka.common.utils.AppInfoParser) -[2023-11-03 19:38:21,249] INFO Kafka commitId: 60e845626d8a465a (org.apache.kafka.common.utils.AppInfoParser) -[2023-11-03 19:38:21,249] INFO Kafka startTimeMs: 1699054701246 (org.apache.kafka.common.utils.AppInfoParser) -[2023-11-03 19:38:21,250] INFO [KafkaServer id=0] started (kafka.server.KafkaServer) -[2023-11-03 19:38:21,398] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Recorded new controller, from now on will use node ThinkPadP53:9092 (id: 0 rack: null) (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:38:21,405] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Recorded new controller, from now on will use node ThinkPadP53:9092 (id: 0 rack: null) (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:38:41,888] INFO Creating topic __consumer_offsets with configuration {compression.type=producer, cleanup.policy=compact, segment.bytes=104857600} and initial partition assignment HashMap(0 -> ArrayBuffer(0), 1 -> ArrayBuffer(0), 2 -> ArrayBuffer(0), 3 -> ArrayBuffer(0), 4 -> ArrayBuffer(0), 5 -> ArrayBuffer(0), 6 -> ArrayBuffer(0), 7 -> ArrayBuffer(0), 8 -> ArrayBuffer(0), 9 -> ArrayBuffer(0), 10 -> ArrayBuffer(0), 11 -> ArrayBuffer(0), 12 -> ArrayBuffer(0), 13 -> ArrayBuffer(0), 14 -> ArrayBuffer(0), 15 -> ArrayBuffer(0), 16 -> ArrayBuffer(0), 17 -> ArrayBuffer(0), 18 -> ArrayBuffer(0), 19 -> ArrayBuffer(0), 20 -> ArrayBuffer(0), 21 -> ArrayBuffer(0), 22 -> ArrayBuffer(0), 23 -> ArrayBuffer(0), 24 -> ArrayBuffer(0), 25 -> ArrayBuffer(0), 26 -> ArrayBuffer(0), 27 -> ArrayBuffer(0), 28 -> ArrayBuffer(0), 29 -> ArrayBuffer(0), 30 -> ArrayBuffer(0), 31 -> ArrayBuffer(0), 32 -> ArrayBuffer(0), 33 -> ArrayBuffer(0), 34 -> ArrayBuffer(0), 35 -> ArrayBuffer(0), 36 -> ArrayBuffer(0), 37 -> ArrayBuffer(0), 38 -> ArrayBuffer(0), 39 -> ArrayBuffer(0), 40 -> ArrayBuffer(0), 41 -> ArrayBuffer(0), 42 -> ArrayBuffer(0), 43 -> ArrayBuffer(0), 44 -> ArrayBuffer(0), 45 -> ArrayBuffer(0), 46 -> ArrayBuffer(0), 47 -> ArrayBuffer(0), 48 -> ArrayBuffer(0), 49 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient) -[2023-11-03 19:38:42,029] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions HashSet(__consumer_offsets-22, __consumer_offsets-30, __consumer_offsets-25, __consumer_offsets-35, __consumer_offsets-37, __consumer_offsets-38, __consumer_offsets-13, __consumer_offsets-8, __consumer_offsets-21, __consumer_offsets-4, __consumer_offsets-27, __consumer_offsets-7, __consumer_offsets-9, __consumer_offsets-46, __consumer_offsets-41, __consumer_offsets-33, __consumer_offsets-23, __consumer_offsets-49, __consumer_offsets-47, __consumer_offsets-16, __consumer_offsets-28, __consumer_offsets-31, __consumer_offsets-36, __consumer_offsets-42, __consumer_offsets-3, __consumer_offsets-18, __consumer_offsets-15, __consumer_offsets-24, __consumer_offsets-17, __consumer_offsets-48, __consumer_offsets-19, __consumer_offsets-11, __consumer_offsets-2, __consumer_offsets-43, __consumer_offsets-6, __consumer_offsets-14, __consumer_offsets-20, __consumer_offsets-0, __consumer_offsets-44, __consumer_offsets-39, __consumer_offsets-12, __consumer_offsets-45, __consumer_offsets-1, __consumer_offsets-5, __consumer_offsets-26, __consumer_offsets-29, __consumer_offsets-34, __consumer_offsets-10, __consumer_offsets-32, __consumer_offsets-40) (kafka.server.ReplicaFetcherManager) -[2023-11-03 19:38:42,067] INFO [LogLoader partition=__consumer_offsets-3, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,077] INFO Created log for partition __consumer_offsets-3 in /tmp/kafka-logs/__consumer_offsets-3 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,079] INFO [Partition __consumer_offsets-3 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-3 (kafka.cluster.Partition) -[2023-11-03 19:38:42,080] INFO [Partition __consumer_offsets-3 broker=0] Log loaded for partition __consumer_offsets-3 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,088] INFO [LogLoader partition=__consumer_offsets-18, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,089] INFO Created log for partition __consumer_offsets-18 in /tmp/kafka-logs/__consumer_offsets-18 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,089] INFO [Partition __consumer_offsets-18 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-18 (kafka.cluster.Partition) -[2023-11-03 19:38:42,089] INFO [Partition __consumer_offsets-18 broker=0] Log loaded for partition __consumer_offsets-18 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,095] INFO [LogLoader partition=__consumer_offsets-41, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,095] INFO Created log for partition __consumer_offsets-41 in /tmp/kafka-logs/__consumer_offsets-41 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,096] INFO [Partition __consumer_offsets-41 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-41 (kafka.cluster.Partition) -[2023-11-03 19:38:42,096] INFO [Partition __consumer_offsets-41 broker=0] Log loaded for partition __consumer_offsets-41 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,099] INFO [LogLoader partition=__consumer_offsets-10, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,100] INFO Created log for partition __consumer_offsets-10 in /tmp/kafka-logs/__consumer_offsets-10 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,100] INFO [Partition __consumer_offsets-10 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-10 (kafka.cluster.Partition) -[2023-11-03 19:38:42,100] INFO [Partition __consumer_offsets-10 broker=0] Log loaded for partition __consumer_offsets-10 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,103] INFO [LogLoader partition=__consumer_offsets-33, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,103] INFO Created log for partition __consumer_offsets-33 in /tmp/kafka-logs/__consumer_offsets-33 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,103] INFO [Partition __consumer_offsets-33 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-33 (kafka.cluster.Partition) -[2023-11-03 19:38:42,104] INFO [Partition __consumer_offsets-33 broker=0] Log loaded for partition __consumer_offsets-33 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,107] INFO [LogLoader partition=__consumer_offsets-48, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,108] INFO Created log for partition __consumer_offsets-48 in /tmp/kafka-logs/__consumer_offsets-48 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,108] INFO [Partition __consumer_offsets-48 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-48 (kafka.cluster.Partition) -[2023-11-03 19:38:42,108] INFO [Partition __consumer_offsets-48 broker=0] Log loaded for partition __consumer_offsets-48 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,113] INFO [LogLoader partition=__consumer_offsets-19, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,114] INFO Created log for partition __consumer_offsets-19 in /tmp/kafka-logs/__consumer_offsets-19 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,114] INFO [Partition __consumer_offsets-19 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-19 (kafka.cluster.Partition) -[2023-11-03 19:38:42,114] INFO [Partition __consumer_offsets-19 broker=0] Log loaded for partition __consumer_offsets-19 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,117] INFO [LogLoader partition=__consumer_offsets-34, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,118] INFO Created log for partition __consumer_offsets-34 in /tmp/kafka-logs/__consumer_offsets-34 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,118] INFO [Partition __consumer_offsets-34 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-34 (kafka.cluster.Partition) -[2023-11-03 19:38:42,118] INFO [Partition __consumer_offsets-34 broker=0] Log loaded for partition __consumer_offsets-34 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,121] INFO [LogLoader partition=__consumer_offsets-4, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,121] INFO Created log for partition __consumer_offsets-4 in /tmp/kafka-logs/__consumer_offsets-4 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,121] INFO [Partition __consumer_offsets-4 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-4 (kafka.cluster.Partition) -[2023-11-03 19:38:42,122] INFO [Partition __consumer_offsets-4 broker=0] Log loaded for partition __consumer_offsets-4 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,127] INFO [LogLoader partition=__consumer_offsets-11, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,127] INFO Created log for partition __consumer_offsets-11 in /tmp/kafka-logs/__consumer_offsets-11 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,127] INFO [Partition __consumer_offsets-11 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-11 (kafka.cluster.Partition) -[2023-11-03 19:38:42,127] INFO [Partition __consumer_offsets-11 broker=0] Log loaded for partition __consumer_offsets-11 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,130] INFO [LogLoader partition=__consumer_offsets-26, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,131] INFO Created log for partition __consumer_offsets-26 in /tmp/kafka-logs/__consumer_offsets-26 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,131] INFO [Partition __consumer_offsets-26 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-26 (kafka.cluster.Partition) -[2023-11-03 19:38:42,131] INFO [Partition __consumer_offsets-26 broker=0] Log loaded for partition __consumer_offsets-26 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,134] INFO [LogLoader partition=__consumer_offsets-49, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,134] INFO Created log for partition __consumer_offsets-49 in /tmp/kafka-logs/__consumer_offsets-49 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,135] INFO [Partition __consumer_offsets-49 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-49 (kafka.cluster.Partition) -[2023-11-03 19:38:42,135] INFO [Partition __consumer_offsets-49 broker=0] Log loaded for partition __consumer_offsets-49 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,137] INFO [LogLoader partition=__consumer_offsets-39, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,137] INFO Created log for partition __consumer_offsets-39 in /tmp/kafka-logs/__consumer_offsets-39 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,137] INFO [Partition __consumer_offsets-39 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-39 (kafka.cluster.Partition) -[2023-11-03 19:38:42,137] INFO [Partition __consumer_offsets-39 broker=0] Log loaded for partition __consumer_offsets-39 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,141] INFO [LogLoader partition=__consumer_offsets-9, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,142] INFO Created log for partition __consumer_offsets-9 in /tmp/kafka-logs/__consumer_offsets-9 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,143] INFO [Partition __consumer_offsets-9 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-9 (kafka.cluster.Partition) -[2023-11-03 19:38:42,143] INFO [Partition __consumer_offsets-9 broker=0] Log loaded for partition __consumer_offsets-9 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,145] INFO [LogLoader partition=__consumer_offsets-24, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,146] INFO Created log for partition __consumer_offsets-24 in /tmp/kafka-logs/__consumer_offsets-24 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,146] INFO [Partition __consumer_offsets-24 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-24 (kafka.cluster.Partition) -[2023-11-03 19:38:42,146] INFO [Partition __consumer_offsets-24 broker=0] Log loaded for partition __consumer_offsets-24 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,149] INFO [LogLoader partition=__consumer_offsets-31, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,149] INFO Created log for partition __consumer_offsets-31 in /tmp/kafka-logs/__consumer_offsets-31 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,149] INFO [Partition __consumer_offsets-31 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-31 (kafka.cluster.Partition) -[2023-11-03 19:38:42,149] INFO [Partition __consumer_offsets-31 broker=0] Log loaded for partition __consumer_offsets-31 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,152] INFO [LogLoader partition=__consumer_offsets-46, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,152] INFO Created log for partition __consumer_offsets-46 in /tmp/kafka-logs/__consumer_offsets-46 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,152] INFO [Partition __consumer_offsets-46 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-46 (kafka.cluster.Partition) -[2023-11-03 19:38:42,152] INFO [Partition __consumer_offsets-46 broker=0] Log loaded for partition __consumer_offsets-46 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,155] INFO [LogLoader partition=__consumer_offsets-1, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,156] INFO Created log for partition __consumer_offsets-1 in /tmp/kafka-logs/__consumer_offsets-1 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,156] INFO [Partition __consumer_offsets-1 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-1 (kafka.cluster.Partition) -[2023-11-03 19:38:42,156] INFO [Partition __consumer_offsets-1 broker=0] Log loaded for partition __consumer_offsets-1 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,162] INFO [LogLoader partition=__consumer_offsets-16, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,163] INFO Created log for partition __consumer_offsets-16 in /tmp/kafka-logs/__consumer_offsets-16 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,163] INFO [Partition __consumer_offsets-16 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-16 (kafka.cluster.Partition) -[2023-11-03 19:38:42,163] INFO [Partition __consumer_offsets-16 broker=0] Log loaded for partition __consumer_offsets-16 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,167] INFO [LogLoader partition=__consumer_offsets-2, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,167] INFO Created log for partition __consumer_offsets-2 in /tmp/kafka-logs/__consumer_offsets-2 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,167] INFO [Partition __consumer_offsets-2 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-2 (kafka.cluster.Partition) -[2023-11-03 19:38:42,167] INFO [Partition __consumer_offsets-2 broker=0] Log loaded for partition __consumer_offsets-2 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,170] INFO [LogLoader partition=__consumer_offsets-25, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,170] INFO Created log for partition __consumer_offsets-25 in /tmp/kafka-logs/__consumer_offsets-25 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,170] INFO [Partition __consumer_offsets-25 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-25 (kafka.cluster.Partition) -[2023-11-03 19:38:42,170] INFO [Partition __consumer_offsets-25 broker=0] Log loaded for partition __consumer_offsets-25 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,173] INFO [LogLoader partition=__consumer_offsets-40, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,175] INFO Created log for partition __consumer_offsets-40 in /tmp/kafka-logs/__consumer_offsets-40 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,175] INFO [Partition __consumer_offsets-40 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-40 (kafka.cluster.Partition) -[2023-11-03 19:38:42,175] INFO [Partition __consumer_offsets-40 broker=0] Log loaded for partition __consumer_offsets-40 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,180] INFO [LogLoader partition=__consumer_offsets-47, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,181] INFO Created log for partition __consumer_offsets-47 in /tmp/kafka-logs/__consumer_offsets-47 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,181] INFO [Partition __consumer_offsets-47 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-47 (kafka.cluster.Partition) -[2023-11-03 19:38:42,181] INFO [Partition __consumer_offsets-47 broker=0] Log loaded for partition __consumer_offsets-47 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,185] INFO [LogLoader partition=__consumer_offsets-17, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,185] INFO Created log for partition __consumer_offsets-17 in /tmp/kafka-logs/__consumer_offsets-17 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,185] INFO [Partition __consumer_offsets-17 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-17 (kafka.cluster.Partition) -[2023-11-03 19:38:42,185] INFO [Partition __consumer_offsets-17 broker=0] Log loaded for partition __consumer_offsets-17 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,188] INFO [LogLoader partition=__consumer_offsets-32, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,188] INFO Created log for partition __consumer_offsets-32 in /tmp/kafka-logs/__consumer_offsets-32 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,188] INFO [Partition __consumer_offsets-32 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-32 (kafka.cluster.Partition) -[2023-11-03 19:38:42,188] INFO [Partition __consumer_offsets-32 broker=0] Log loaded for partition __consumer_offsets-32 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,194] INFO [LogLoader partition=__consumer_offsets-37, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,195] INFO Created log for partition __consumer_offsets-37 in /tmp/kafka-logs/__consumer_offsets-37 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,195] INFO [Partition __consumer_offsets-37 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-37 (kafka.cluster.Partition) -[2023-11-03 19:38:42,195] INFO [Partition __consumer_offsets-37 broker=0] Log loaded for partition __consumer_offsets-37 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,200] INFO [LogLoader partition=__consumer_offsets-7, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,201] INFO Created log for partition __consumer_offsets-7 in /tmp/kafka-logs/__consumer_offsets-7 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,201] INFO [Partition __consumer_offsets-7 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-7 (kafka.cluster.Partition) -[2023-11-03 19:38:42,201] INFO [Partition __consumer_offsets-7 broker=0] Log loaded for partition __consumer_offsets-7 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,204] INFO [LogLoader partition=__consumer_offsets-22, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,204] INFO Created log for partition __consumer_offsets-22 in /tmp/kafka-logs/__consumer_offsets-22 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,204] INFO [Partition __consumer_offsets-22 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-22 (kafka.cluster.Partition) -[2023-11-03 19:38:42,204] INFO [Partition __consumer_offsets-22 broker=0] Log loaded for partition __consumer_offsets-22 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,209] INFO [LogLoader partition=__consumer_offsets-29, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,210] INFO Created log for partition __consumer_offsets-29 in /tmp/kafka-logs/__consumer_offsets-29 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,210] INFO [Partition __consumer_offsets-29 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-29 (kafka.cluster.Partition) -[2023-11-03 19:38:42,210] INFO [Partition __consumer_offsets-29 broker=0] Log loaded for partition __consumer_offsets-29 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,215] INFO [LogLoader partition=__consumer_offsets-44, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,216] INFO Created log for partition __consumer_offsets-44 in /tmp/kafka-logs/__consumer_offsets-44 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,216] INFO [Partition __consumer_offsets-44 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-44 (kafka.cluster.Partition) -[2023-11-03 19:38:42,216] INFO [Partition __consumer_offsets-44 broker=0] Log loaded for partition __consumer_offsets-44 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,220] INFO [LogLoader partition=__consumer_offsets-14, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,220] INFO Created log for partition __consumer_offsets-14 in /tmp/kafka-logs/__consumer_offsets-14 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,220] INFO [Partition __consumer_offsets-14 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-14 (kafka.cluster.Partition) -[2023-11-03 19:38:42,220] INFO [Partition __consumer_offsets-14 broker=0] Log loaded for partition __consumer_offsets-14 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,225] INFO [LogLoader partition=__consumer_offsets-23, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,226] INFO Created log for partition __consumer_offsets-23 in /tmp/kafka-logs/__consumer_offsets-23 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,226] INFO [Partition __consumer_offsets-23 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-23 (kafka.cluster.Partition) -[2023-11-03 19:38:42,226] INFO [Partition __consumer_offsets-23 broker=0] Log loaded for partition __consumer_offsets-23 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,230] INFO [LogLoader partition=__consumer_offsets-38, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,231] INFO Created log for partition __consumer_offsets-38 in /tmp/kafka-logs/__consumer_offsets-38 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,231] INFO [Partition __consumer_offsets-38 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-38 (kafka.cluster.Partition) -[2023-11-03 19:38:42,231] INFO [Partition __consumer_offsets-38 broker=0] Log loaded for partition __consumer_offsets-38 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,235] INFO [LogLoader partition=__consumer_offsets-8, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,236] INFO Created log for partition __consumer_offsets-8 in /tmp/kafka-logs/__consumer_offsets-8 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,236] INFO [Partition __consumer_offsets-8 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-8 (kafka.cluster.Partition) -[2023-11-03 19:38:42,236] INFO [Partition __consumer_offsets-8 broker=0] Log loaded for partition __consumer_offsets-8 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,238] INFO [LogLoader partition=__consumer_offsets-45, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,239] INFO Created log for partition __consumer_offsets-45 in /tmp/kafka-logs/__consumer_offsets-45 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,239] INFO [Partition __consumer_offsets-45 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-45 (kafka.cluster.Partition) -[2023-11-03 19:38:42,239] INFO [Partition __consumer_offsets-45 broker=0] Log loaded for partition __consumer_offsets-45 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,244] INFO [LogLoader partition=__consumer_offsets-15, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,244] INFO Created log for partition __consumer_offsets-15 in /tmp/kafka-logs/__consumer_offsets-15 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,244] INFO [Partition __consumer_offsets-15 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-15 (kafka.cluster.Partition) -[2023-11-03 19:38:42,244] INFO [Partition __consumer_offsets-15 broker=0] Log loaded for partition __consumer_offsets-15 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,248] INFO [LogLoader partition=__consumer_offsets-30, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,249] INFO Created log for partition __consumer_offsets-30 in /tmp/kafka-logs/__consumer_offsets-30 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,249] INFO [Partition __consumer_offsets-30 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-30 (kafka.cluster.Partition) -[2023-11-03 19:38:42,249] INFO [Partition __consumer_offsets-30 broker=0] Log loaded for partition __consumer_offsets-30 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,251] INFO [LogLoader partition=__consumer_offsets-0, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,251] INFO Created log for partition __consumer_offsets-0 in /tmp/kafka-logs/__consumer_offsets-0 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,252] INFO [Partition __consumer_offsets-0 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,252] INFO [Partition __consumer_offsets-0 broker=0] Log loaded for partition __consumer_offsets-0 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,254] INFO [LogLoader partition=__consumer_offsets-35, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,254] INFO Created log for partition __consumer_offsets-35 in /tmp/kafka-logs/__consumer_offsets-35 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,254] INFO [Partition __consumer_offsets-35 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-35 (kafka.cluster.Partition) -[2023-11-03 19:38:42,254] INFO [Partition __consumer_offsets-35 broker=0] Log loaded for partition __consumer_offsets-35 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,260] INFO [LogLoader partition=__consumer_offsets-5, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,260] INFO Created log for partition __consumer_offsets-5 in /tmp/kafka-logs/__consumer_offsets-5 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,261] INFO [Partition __consumer_offsets-5 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-5 (kafka.cluster.Partition) -[2023-11-03 19:38:42,261] INFO [Partition __consumer_offsets-5 broker=0] Log loaded for partition __consumer_offsets-5 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,264] INFO [LogLoader partition=__consumer_offsets-20, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,265] INFO Created log for partition __consumer_offsets-20 in /tmp/kafka-logs/__consumer_offsets-20 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,265] INFO [Partition __consumer_offsets-20 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-20 (kafka.cluster.Partition) -[2023-11-03 19:38:42,266] INFO [Partition __consumer_offsets-20 broker=0] Log loaded for partition __consumer_offsets-20 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,269] INFO [LogLoader partition=__consumer_offsets-27, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,269] INFO Created log for partition __consumer_offsets-27 in /tmp/kafka-logs/__consumer_offsets-27 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,269] INFO [Partition __consumer_offsets-27 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-27 (kafka.cluster.Partition) -[2023-11-03 19:38:42,270] INFO [Partition __consumer_offsets-27 broker=0] Log loaded for partition __consumer_offsets-27 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,272] INFO [LogLoader partition=__consumer_offsets-42, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,273] INFO Created log for partition __consumer_offsets-42 in /tmp/kafka-logs/__consumer_offsets-42 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,273] INFO [Partition __consumer_offsets-42 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-42 (kafka.cluster.Partition) -[2023-11-03 19:38:42,273] INFO [Partition __consumer_offsets-42 broker=0] Log loaded for partition __consumer_offsets-42 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,277] INFO [LogLoader partition=__consumer_offsets-12, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,277] INFO Created log for partition __consumer_offsets-12 in /tmp/kafka-logs/__consumer_offsets-12 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,277] INFO [Partition __consumer_offsets-12 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-12 (kafka.cluster.Partition) -[2023-11-03 19:38:42,277] INFO [Partition __consumer_offsets-12 broker=0] Log loaded for partition __consumer_offsets-12 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,280] INFO [LogLoader partition=__consumer_offsets-21, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,280] INFO Created log for partition __consumer_offsets-21 in /tmp/kafka-logs/__consumer_offsets-21 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,280] INFO [Partition __consumer_offsets-21 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-21 (kafka.cluster.Partition) -[2023-11-03 19:38:42,280] INFO [Partition __consumer_offsets-21 broker=0] Log loaded for partition __consumer_offsets-21 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,284] INFO [LogLoader partition=__consumer_offsets-36, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,285] INFO Created log for partition __consumer_offsets-36 in /tmp/kafka-logs/__consumer_offsets-36 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,285] INFO [Partition __consumer_offsets-36 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-36 (kafka.cluster.Partition) -[2023-11-03 19:38:42,285] INFO [Partition __consumer_offsets-36 broker=0] Log loaded for partition __consumer_offsets-36 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,288] INFO [LogLoader partition=__consumer_offsets-6, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,288] INFO Created log for partition __consumer_offsets-6 in /tmp/kafka-logs/__consumer_offsets-6 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,288] INFO [Partition __consumer_offsets-6 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-6 (kafka.cluster.Partition) -[2023-11-03 19:38:42,288] INFO [Partition __consumer_offsets-6 broker=0] Log loaded for partition __consumer_offsets-6 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,291] INFO [LogLoader partition=__consumer_offsets-43, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,292] INFO Created log for partition __consumer_offsets-43 in /tmp/kafka-logs/__consumer_offsets-43 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,292] INFO [Partition __consumer_offsets-43 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-43 (kafka.cluster.Partition) -[2023-11-03 19:38:42,292] INFO [Partition __consumer_offsets-43 broker=0] Log loaded for partition __consumer_offsets-43 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,295] INFO [LogLoader partition=__consumer_offsets-13, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,295] INFO Created log for partition __consumer_offsets-13 in /tmp/kafka-logs/__consumer_offsets-13 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,295] INFO [Partition __consumer_offsets-13 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-13 (kafka.cluster.Partition) -[2023-11-03 19:38:42,296] INFO [Partition __consumer_offsets-13 broker=0] Log loaded for partition __consumer_offsets-13 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,298] INFO [LogLoader partition=__consumer_offsets-28, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:42,299] INFO Created log for partition __consumer_offsets-28 in /tmp/kafka-logs/__consumer_offsets-28 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) -[2023-11-03 19:38:42,299] INFO [Partition __consumer_offsets-28 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-28 (kafka.cluster.Partition) -[2023-11-03 19:38:42,299] INFO [Partition __consumer_offsets-28 broker=0] Log loaded for partition __consumer_offsets-28 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:38:42,303] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 3 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,303] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-3 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 18 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-18 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 41 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-41 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 10 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-10 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 33 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-33 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 48 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-48 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 19 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-19 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 34 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-34 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 4 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-4 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 11 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-11 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 26 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-26 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 49 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-49 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 39 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-39 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 9 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-9 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 24 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-24 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 31 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,304] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-31 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,304] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 46 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-46 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 1 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-1 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 16 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-16 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 2 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-2 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 25 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-25 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 40 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-40 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 47 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-47 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 17 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-17 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 32 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-32 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 37 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-37 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 7 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-7 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 22 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-22 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 29 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-29 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 44 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-44 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 14 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-14 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 23 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-23 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 38 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-38 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 8 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-8 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 45 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-45 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 15 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-15 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 30 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-30 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 0 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,305] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-0 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,305] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 35 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-35 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 5 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-5 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 20 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-20 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 27 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-27 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 42 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-42 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 12 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-12 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 21 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-21 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 36 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-36 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 6 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-6 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 43 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-43 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 13 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-13 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,306] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 28 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:38:42,306] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-28 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,310] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-3 in 4 milliseconds for epoch 0, of which 1 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,310] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-18 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,310] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-41 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,311] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-10 in 7 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,311] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-33 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,311] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-48 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,311] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-19 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,311] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-34 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,311] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-4 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,311] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-11 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,311] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-26 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,311] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-49 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,312] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-39 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,312] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-9 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,312] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-24 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,312] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-31 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,312] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-46 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,312] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-1 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,312] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-16 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,312] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-2 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-25 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-40 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-47 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-17 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-32 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-37 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-7 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-22 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-29 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-44 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,313] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-14 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-23 in 9 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-38 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-8 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-45 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-15 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-30 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-0 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-35 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-5 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-20 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-27 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-42 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-12 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-21 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,314] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-36 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,315] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-6 in 8 milliseconds for epoch 0, of which 8 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,315] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-43 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,315] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-13 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:42,315] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-28 in 9 milliseconds for epoch 0, of which 9 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:38:56,514] INFO Creating topic test-topic with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient) -[2023-11-03 19:38:56,527] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(test-topic-0) (kafka.server.ReplicaFetcherManager) -[2023-11-03 19:38:56,529] INFO [LogLoader partition=test-topic-0, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:38:56,529] INFO Created log for partition test-topic-0 in /tmp/kafka-logs/test-topic-0 with properties {} (kafka.log.LogManager) -[2023-11-03 19:38:56,530] INFO [Partition test-topic-0 broker=0] No checkpointed highwatermark is found for partition test-topic-0 (kafka.cluster.Partition) -[2023-11-03 19:38:56,530] INFO [Partition test-topic-0 broker=0] Log loaded for partition test-topic-0 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:39:07,767] INFO [GroupCoordinator 0]: Dynamic member with unknown member id joins group test-group in Empty state. Created a new member id rdkafka-d8d257cb-90fd-4aed-a191-0c8705f20f1a and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:39:07,775] INFO [GroupCoordinator 0]: Preparing to rebalance group test-group in state PreparingRebalance with old generation 0 (__consumer_offsets-12) (reason: Adding new member rdkafka-d8d257cb-90fd-4aed-a191-0c8705f20f1a with group instance id None; client reason: not provided) (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:39:07,791] INFO [GroupCoordinator 0]: Stabilized group test-group generation 1 (__consumer_offsets-12) with 1 members (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:39:07,797] INFO [GroupCoordinator 0]: Assignment received from leader rdkafka-d8d257cb-90fd-4aed-a191-0c8705f20f1a for group test-group for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:40:46,816] INFO [GroupCoordinator 0]: Member rdkafka-d8d257cb-90fd-4aed-a191-0c8705f20f1a in group test-group has failed, removing it from the group (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:40:46,817] INFO [GroupCoordinator 0]: Preparing to rebalance group test-group in state PreparingRebalance with old generation 1 (__consumer_offsets-12) (reason: removing member rdkafka-d8d257cb-90fd-4aed-a191-0c8705f20f1a on heartbeat expiration) (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:40:46,818] INFO [GroupCoordinator 0]: Group test-group with generation 2 is now empty (__consumer_offsets-12) (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:40:48,852] INFO Terminating process due to signal SIGTERM (org.apache.kafka.common.utils.LoggingSignalHandler) -[2023-11-03 19:40:48,853] INFO [KafkaServer id=0] shutting down (kafka.server.KafkaServer) -[2023-11-03 19:40:48,854] INFO [KafkaServer id=0] Starting controlled shutdown (kafka.server.KafkaServer) -[2023-11-03 19:40:48,865] INFO [KafkaServer id=0] Controlled shutdown request returned successfully after 5ms (kafka.server.KafkaServer) -[2023-11-03 19:40:48,867] INFO [/config/changes-event-process-thread]: Shutting down (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread) -[2023-11-03 19:40:48,867] INFO [/config/changes-event-process-thread]: Stopped (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread) -[2023-11-03 19:40:48,867] INFO [/config/changes-event-process-thread]: Shutdown completed (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread) -[2023-11-03 19:40:48,868] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Stopping socket server request processors (kafka.network.SocketServer) -[2023-11-03 19:40:48,875] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Stopped socket server request processors (kafka.network.SocketServer) -[2023-11-03 19:40:48,876] INFO [data-plane Kafka Request Handler on Broker 0], shutting down (kafka.server.KafkaRequestHandlerPool) -[2023-11-03 19:40:48,876] INFO [data-plane Kafka Request Handler on Broker 0], shut down completely (kafka.server.KafkaRequestHandlerPool) -[2023-11-03 19:40:48,878] INFO [ExpirationReaper-0-AlterAcls]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,878] INFO [ExpirationReaper-0-AlterAcls]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,878] INFO [ExpirationReaper-0-AlterAcls]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,879] INFO [KafkaApi-0] Shutdown complete. (kafka.server.KafkaApis) -[2023-11-03 19:40:48,879] INFO [ExpirationReaper-0-topic]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,879] INFO [ExpirationReaper-0-topic]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,879] INFO [ExpirationReaper-0-topic]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,880] INFO [TransactionCoordinator id=0] Shutting down. (kafka.coordinator.transaction.TransactionCoordinator) -[2023-11-03 19:40:48,880] INFO [Transaction State Manager 0]: Shutdown complete (kafka.coordinator.transaction.TransactionStateManager) -[2023-11-03 19:40:48,880] INFO [TxnMarkerSenderThread-0]: Shutting down (kafka.coordinator.transaction.TransactionMarkerChannelManager) -[2023-11-03 19:40:48,881] INFO [TxnMarkerSenderThread-0]: Stopped (kafka.coordinator.transaction.TransactionMarkerChannelManager) -[2023-11-03 19:40:48,881] INFO [TxnMarkerSenderThread-0]: Shutdown completed (kafka.coordinator.transaction.TransactionMarkerChannelManager) -[2023-11-03 19:40:48,881] INFO [TransactionCoordinator id=0] Shutdown complete. (kafka.coordinator.transaction.TransactionCoordinator) -[2023-11-03 19:40:48,881] INFO [GroupCoordinator 0]: Shutting down. (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:40:48,882] INFO [ExpirationReaper-0-Heartbeat]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,882] INFO [ExpirationReaper-0-Heartbeat]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,882] INFO [ExpirationReaper-0-Heartbeat]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,882] INFO [ExpirationReaper-0-Rebalance]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,882] INFO [ExpirationReaper-0-Rebalance]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,882] INFO [ExpirationReaper-0-Rebalance]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,882] INFO [GroupCoordinator 0]: Shutdown complete. (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:40:48,883] INFO [ReplicaManager broker=0] Shutting down (kafka.server.ReplicaManager) -[2023-11-03 19:40:48,883] INFO [LogDirFailureHandler]: Shutting down (kafka.server.ReplicaManager$LogDirFailureHandler) -[2023-11-03 19:40:48,883] INFO [LogDirFailureHandler]: Stopped (kafka.server.ReplicaManager$LogDirFailureHandler) -[2023-11-03 19:40:48,883] INFO [LogDirFailureHandler]: Shutdown completed (kafka.server.ReplicaManager$LogDirFailureHandler) -[2023-11-03 19:40:48,884] INFO [ReplicaFetcherManager on broker 0] shutting down (kafka.server.ReplicaFetcherManager) -[2023-11-03 19:40:48,884] INFO [ReplicaFetcherManager on broker 0] shutdown completed (kafka.server.ReplicaFetcherManager) -[2023-11-03 19:40:48,884] INFO [ReplicaAlterLogDirsManager on broker 0] shutting down (kafka.server.ReplicaAlterLogDirsManager) -[2023-11-03 19:40:48,884] INFO [ReplicaAlterLogDirsManager on broker 0] shutdown completed (kafka.server.ReplicaAlterLogDirsManager) -[2023-11-03 19:40:48,884] INFO [ExpirationReaper-0-Fetch]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,884] INFO [ExpirationReaper-0-Fetch]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,884] INFO [ExpirationReaper-0-Fetch]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,885] INFO [ExpirationReaper-0-RemoteFetch]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,885] INFO [ExpirationReaper-0-RemoteFetch]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,885] INFO [ExpirationReaper-0-RemoteFetch]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,885] INFO [ExpirationReaper-0-Produce]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,885] INFO [ExpirationReaper-0-Produce]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,885] INFO [ExpirationReaper-0-Produce]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,885] INFO [ExpirationReaper-0-DeleteRecords]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,885] INFO [ExpirationReaper-0-DeleteRecords]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,886] INFO [ExpirationReaper-0-DeleteRecords]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,886] INFO [ExpirationReaper-0-ElectLeader]: Shutting down (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,886] INFO [ExpirationReaper-0-ElectLeader]: Stopped (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,886] INFO [ExpirationReaper-0-ElectLeader]: Shutdown completed (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:40:48,889] INFO [AddPartitionsToTxnSenderThread-0]: Shutting down (kafka.server.AddPartitionsToTxnManager) -[2023-11-03 19:40:48,889] INFO [AddPartitionsToTxnSenderThread-0]: Stopped (kafka.server.AddPartitionsToTxnManager) -[2023-11-03 19:40:48,889] INFO [AddPartitionsToTxnSenderThread-0]: Shutdown completed (kafka.server.AddPartitionsToTxnManager) -[2023-11-03 19:40:48,889] INFO [ReplicaManager broker=0] Shut down completely (kafka.server.ReplicaManager) -[2023-11-03 19:40:48,890] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Shutting down (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:40:48,890] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Stopped (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:40:48,890] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Shutdown completed (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:40:48,892] INFO Broker to controller channel manager for alter-partition shutdown (kafka.server.BrokerToControllerChannelManagerImpl) -[2023-11-03 19:40:48,892] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Shutting down (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:40:48,893] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Stopped (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:40:48,893] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Shutdown completed (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:40:48,893] INFO Broker to controller channel manager for forwarding shutdown (kafka.server.BrokerToControllerChannelManagerImpl) -[2023-11-03 19:40:48,894] INFO Shutting down. (kafka.log.LogManager) -[2023-11-03 19:40:48,894] INFO [kafka-log-cleaner-thread-0]: Shutting down (kafka.log.LogCleaner$CleanerThread) -[2023-11-03 19:40:48,894] INFO [kafka-log-cleaner-thread-0]: Stopped (kafka.log.LogCleaner$CleanerThread) -[2023-11-03 19:40:48,894] INFO [kafka-log-cleaner-thread-0]: Shutdown completed (kafka.log.LogCleaner$CleanerThread) -[2023-11-03 19:40:48,912] INFO [ProducerStateManager partition=__consumer_offsets-12]Wrote producer snapshot at offset 7 with 0 producer ids in 0 ms. (org.apache.kafka.storage.internals.log.ProducerStateManager) -[2023-11-03 19:40:48,914] INFO [ProducerStateManager partition=test-topic-0]Wrote producer snapshot at offset 5 with 0 producer ids in 0 ms. (org.apache.kafka.storage.internals.log.ProducerStateManager) -[2023-11-03 19:40:48,920] INFO Shutdown complete. (kafka.log.LogManager) -[2023-11-03 19:40:48,925] INFO [feature-zk-node-event-process-thread]: Shutting down (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) -[2023-11-03 19:40:48,925] INFO [feature-zk-node-event-process-thread]: Stopped (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) -[2023-11-03 19:40:48,925] INFO [feature-zk-node-event-process-thread]: Shutdown completed (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) -[2023-11-03 19:40:48,925] INFO [ZooKeeperClient Kafka server] Closing. (kafka.zookeeper.ZooKeeperClient) -[2023-11-03 19:40:49,028] INFO Session: 0x100008497380000 closed (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:40:49,028] INFO EventThread shut down for session: 0x100008497380000 (org.apache.zookeeper.ClientCnxn) -[2023-11-03 19:40:49,028] INFO [ZooKeeperClient Kafka server] Closed. (kafka.zookeeper.ZooKeeperClient) -[2023-11-03 19:40:49,028] INFO [ThrottledChannelReaper-Fetch]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-Fetch]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-Fetch]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-Produce]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-Produce]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-Produce]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-Request]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-Request]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-Request]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-ControllerMutation]: Shutting down (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-ControllerMutation]: Stopped (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,029] INFO [ThrottledChannelReaper-ControllerMutation]: Shutdown completed (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:40:49,030] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Shutting down socket server (kafka.network.SocketServer) -[2023-11-03 19:40:49,039] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Shutdown completed (kafka.network.SocketServer) -[2023-11-03 19:40:49,039] INFO Metrics scheduler closed (org.apache.kafka.common.metrics.Metrics) -[2023-11-03 19:40:49,039] INFO Closing reporter org.apache.kafka.common.metrics.JmxReporter (org.apache.kafka.common.metrics.Metrics) -[2023-11-03 19:40:49,039] INFO Metrics reporters closed (org.apache.kafka.common.metrics.Metrics) -[2023-11-03 19:40:49,040] INFO Broker and topic stats closed (kafka.server.BrokerTopicStats) -[2023-11-03 19:40:49,040] INFO App info kafka.server for 0 unregistered (org.apache.kafka.common.utils.AppInfoParser) -[2023-11-03 19:40:49,040] INFO [KafkaServer id=0] shut down completed (kafka.server.KafkaServer) -[2023-11-03 19:45:15,871] INFO Reading configuration from: ./config/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:45:15,874] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:45:15,874] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:45:15,874] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:45:15,874] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:45:15,877] INFO autopurge.snapRetainCount set to 3 (org.apache.zookeeper.server.DatadirCleanupManager) -[2023-11-03 19:45:15,877] INFO autopurge.purgeInterval set to 0 (org.apache.zookeeper.server.DatadirCleanupManager) -[2023-11-03 19:45:15,877] INFO Purge task is not scheduled. (org.apache.zookeeper.server.DatadirCleanupManager) -[2023-11-03 19:45:15,877] WARN Either no config or no quorum defined in config, running in standalone mode (org.apache.zookeeper.server.quorum.QuorumPeerMain) -[2023-11-03 19:45:15,878] INFO Log4j 1.2 jmx support not found; jmx disabled. (org.apache.zookeeper.jmx.ManagedUtil) -[2023-11-03 19:45:15,879] INFO Reading configuration from: ./config/zookeeper.properties (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:45:15,879] INFO clientPortAddress is 0.0.0.0:2181 (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:45:15,879] INFO secureClientPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:45:15,879] INFO observerMasterPort is not set (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:45:15,879] INFO metricsProvider.className is org.apache.zookeeper.metrics.impl.DefaultMetricsProvider (org.apache.zookeeper.server.quorum.QuorumPeerConfig) -[2023-11-03 19:45:15,879] INFO Starting server (org.apache.zookeeper.server.ZooKeeperServerMain) -[2023-11-03 19:45:15,887] INFO ServerMetrics initialized with provider org.apache.zookeeper.metrics.impl.DefaultMetricsProvider@4034c28c (org.apache.zookeeper.server.ServerMetrics) -[2023-11-03 19:45:15,888] INFO ACL digest algorithm is: SHA1 (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) -[2023-11-03 19:45:15,888] INFO zookeeper.DigestAuthenticationProvider.enabled = true (org.apache.zookeeper.server.auth.DigestAuthenticationProvider) -[2023-11-03 19:45:15,891] INFO zookeeper.snapshot.trust.empty : false (org.apache.zookeeper.server.persistence.FileTxnSnapLog) -[2023-11-03 19:45:15,899] INFO (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,899] INFO ______ _ (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,899] INFO |___ / | | (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,899] INFO / / ___ ___ | | __ ___ ___ _ __ ___ _ __ (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,899] INFO / / / _ \ / _ \ | |/ / / _ \ / _ \ | '_ \ / _ \ | '__| (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,899] INFO / /__ | (_) | | (_) | | < | __/ | __/ | |_) | | __/ | | (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,899] INFO /_____| \___/ \___/ |_|\_\ \___| \___| | .__/ \___| |_| (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,900] INFO | | (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,900] INFO |_| (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,900] INFO (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,902] INFO Server environment:zookeeper.version=3.8.2-139d619b58292d7734b4fc83a0f44be4e7b0c986, built on 2023-07-05 19:24 UTC (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,902] INFO Server environment:host.name=ThinkPadP53 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,902] INFO Server environment:java.version=17.0.6 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,902] INFO Server environment:java.vendor=Eclipse Adoptium (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,902] INFO Server environment:java.home=/opt/openjdk-bin-17.0.6_p10 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,902] INFO Server environment:java.class.path=/scratch/Exemple_Kafka/bin/../libs/activation-1.1.1.jar:/scratch/Exemple_Kafka/bin/../libs/aopalliance-repackaged-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/argparse4j-0.7.0.jar:/scratch/Exemple_Kafka/bin/../libs/audience-annotations-0.12.0.jar:/scratch/Exemple_Kafka/bin/../libs/caffeine-2.9.3.jar:/scratch/Exemple_Kafka/bin/../libs/checker-qual-3.19.0.jar:/scratch/Exemple_Kafka/bin/../libs/commons-beanutils-1.9.4.jar:/scratch/Exemple_Kafka/bin/../libs/commons-cli-1.4.jar:/scratch/Exemple_Kafka/bin/../libs/commons-collections-3.2.2.jar:/scratch/Exemple_Kafka/bin/../libs/commons-digester-2.1.jar:/scratch/Exemple_Kafka/bin/../libs/commons-io-2.11.0.jar:/scratch/Exemple_Kafka/bin/../libs/commons-lang3-3.8.1.jar:/scratch/Exemple_Kafka/bin/../libs/commons-logging-1.2.jar:/scratch/Exemple_Kafka/bin/../libs/commons-validator-1.7.jar:/scratch/Exemple_Kafka/bin/../libs/connect-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-basic-auth-extension-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-json-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-mirror-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-mirror-client-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-runtime-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-transforms-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/error_prone_annotations-2.10.0.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-api-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-locator-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-utils-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-annotations-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-core-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-databind-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-dataformat-csv-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-datatype-jdk8-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-jaxrs-base-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-jaxrs-json-provider-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-module-jaxb-annotations-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-module-scala_2.13-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.activation-api-1.2.2.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.annotation-api-1.3.5.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.inject-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.validation-api-2.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.ws.rs-api-2.1.6.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.xml.bind-api-2.3.3.jar:/scratch/Exemple_Kafka/bin/../libs/javassist-3.29.2-GA.jar:/scratch/Exemple_Kafka/bin/../libs/javax.activation-api-1.2.0.jar:/scratch/Exemple_Kafka/bin/../libs/javax.annotation-api-1.3.2.jar:/scratch/Exemple_Kafka/bin/../libs/javax.servlet-api-3.1.0.jar:/scratch/Exemple_Kafka/bin/../libs/javax.ws.rs-api-2.1.1.jar:/scratch/Exemple_Kafka/bin/../libs/jaxb-api-2.3.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-client-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-common-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-container-servlet-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-container-servlet-core-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-hk2-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-server-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-client-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-continuation-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-http-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-io-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-security-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-server-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-servlet-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-servlets-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-util-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-util-ajax-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jline-3.22.0.jar:/scratch/Exemple_Kafka/bin/../libs/jopt-simple-5.0.4.jar:/scratch/Exemple_Kafka/bin/../libs/jose4j-0.9.3.jar:/scratch/Exemple_Kafka/bin/../libs/jsr305-3.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-clients-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-group-coordinator-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-log4j-appender-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-metadata-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-raft-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-server-common-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-shell-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-storage-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-storage-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-examples-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-scala_2.13-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-test-utils-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-tools-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-tools-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka_2.13-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/lz4-java-1.8.0.jar:/scratch/Exemple_Kafka/bin/../libs/maven-artifact-3.8.8.jar:/scratch/Exemple_Kafka/bin/../libs/metrics-core-2.2.0.jar:/scratch/Exemple_Kafka/bin/../libs/metrics-core-4.1.12.1.jar:/scratch/Exemple_Kafka/bin/../libs/netty-buffer-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-codec-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-common-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-handler-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-resolver-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-classes-epoll-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-native-epoll-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-native-unix-common-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/osgi-resource-locator-1.0.3.jar:/scratch/Exemple_Kafka/bin/../libs/paranamer-2.8.jar:/scratch/Exemple_Kafka/bin/../libs/pcollections-4.0.1.jar:/scratch/Exemple_Kafka/bin/../libs/plexus-utils-3.3.1.jar:/scratch/Exemple_Kafka/bin/../libs/reflections-0.10.2.jar:/scratch/Exemple_Kafka/bin/../libs/reload4j-1.2.25.jar:/scratch/Exemple_Kafka/bin/../libs/rocksdbjni-7.9.2.jar:/scratch/Exemple_Kafka/bin/../libs/scala-collection-compat_2.13-2.10.0.jar:/scratch/Exemple_Kafka/bin/../libs/scala-java8-compat_2.13-1.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/scala-library-2.13.11.jar:/scratch/Exemple_Kafka/bin/../libs/scala-logging_2.13-3.9.4.jar:/scratch/Exemple_Kafka/bin/../libs/scala-reflect-2.13.11.jar:/scratch/Exemple_Kafka/bin/../libs/slf4j-api-1.7.36.jar:/scratch/Exemple_Kafka/bin/../libs/slf4j-reload4j-1.7.36.jar:/scratch/Exemple_Kafka/bin/../libs/snappy-java-1.1.10.4.jar:/scratch/Exemple_Kafka/bin/../libs/swagger-annotations-2.2.8.jar:/scratch/Exemple_Kafka/bin/../libs/trogdor-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/zookeeper-3.8.2.jar:/scratch/Exemple_Kafka/bin/../libs/zookeeper-jute-3.8.2.jar:/scratch/Exemple_Kafka/bin/../libs/zstd-jni-1.5.5-1.jar (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,902] INFO Server environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,902] INFO Server environment:java.io.tmpdir=/tmp (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,902] INFO Server environment:java.compiler= (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO Server environment:os.name=Linux (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO Server environment:os.arch=amd64 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO Server environment:os.version=6.4.3-cachyosGentooThinkPadP53 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO Server environment:user.name=memartel (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO Server environment:user.home=/home/memartel (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO Server environment:user.dir=/scratch/Exemple_Kafka (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO Server environment:os.memory.free=494MB (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO Server environment:os.memory.max=512MB (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO Server environment:os.memory.total=512MB (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO zookeeper.enableEagerACLCheck = false (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO zookeeper.digest.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO zookeeper.closeSessionTxn.enabled = true (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO zookeeper.flushDelay = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO zookeeper.maxWriteQueuePollTime = 0 ms (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO zookeeper.maxBatchSize=1000 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,903] INFO zookeeper.intBufferStartingSizeBytes = 1024 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,904] INFO Weighed connection throttling is disabled (org.apache.zookeeper.server.BlueThrottle) -[2023-11-03 19:45:15,905] INFO minSessionTimeout set to 6000 ms (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,905] INFO maxSessionTimeout set to 60000 ms (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,906] INFO getData response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) -[2023-11-03 19:45:15,906] INFO getChildren response cache size is initialized with value 400. (org.apache.zookeeper.server.ResponseCache) -[2023-11-03 19:45:15,907] INFO zookeeper.pathStats.slotCapacity = 60 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:45:15,907] INFO zookeeper.pathStats.slotDuration = 15 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:45:15,907] INFO zookeeper.pathStats.maxDepth = 6 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:45:15,907] INFO zookeeper.pathStats.initialDelay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:45:15,907] INFO zookeeper.pathStats.delay = 5 (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:45:15,907] INFO zookeeper.pathStats.enabled = false (org.apache.zookeeper.server.util.RequestPathMetricsCollector) -[2023-11-03 19:45:15,909] INFO The max bytes for all large requests are set to 104857600 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,910] INFO The large request threshold is set to -1 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,911] INFO zookeeper.enforce.auth.enabled = false (org.apache.zookeeper.server.AuthenticationHelper) -[2023-11-03 19:45:15,911] INFO zookeeper.enforce.auth.schemes = [] (org.apache.zookeeper.server.AuthenticationHelper) -[2023-11-03 19:45:15,911] INFO Created server with tickTime 3000 ms minSessionTimeout 6000 ms maxSessionTimeout 60000 ms clientPortListenBacklog -1 datadir /tmp/zookeeper/version-2 snapdir /tmp/zookeeper/version-2 (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,916] INFO Using org.apache.zookeeper.server.NIOServerCnxnFactory as server connection factory (org.apache.zookeeper.server.ServerCnxnFactory) -[2023-11-03 19:45:15,917] WARN maxCnxns is not configured, using default value 0. (org.apache.zookeeper.server.ServerCnxnFactory) -[2023-11-03 19:45:15,918] INFO Configuring NIO connection handler with 10s sessionless connection timeout, 2 selector thread(s), 24 worker threads, and 64 kB direct buffers. (org.apache.zookeeper.server.NIOServerCnxnFactory) -[2023-11-03 19:45:15,922] INFO binding to port 0.0.0.0/0.0.0.0:2181 (org.apache.zookeeper.server.NIOServerCnxnFactory) -[2023-11-03 19:45:15,935] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) -[2023-11-03 19:45:15,936] INFO Using org.apache.zookeeper.server.watch.WatchManager as watch manager (org.apache.zookeeper.server.watch.WatchManagerFactory) -[2023-11-03 19:45:15,936] INFO zookeeper.snapshotSizeFactor = 0.33 (org.apache.zookeeper.server.ZKDatabase) -[2023-11-03 19:45:15,936] INFO zookeeper.commitLogCount=500 (org.apache.zookeeper.server.ZKDatabase) -[2023-11-03 19:45:15,938] INFO zookeeper.snapshot.compression.method = CHECKED (org.apache.zookeeper.server.persistence.SnapStream) -[2023-11-03 19:45:15,938] INFO Reading snapshot /tmp/zookeeper/version-2/snapshot.0 (org.apache.zookeeper.server.persistence.FileSnap) -[2023-11-03 19:45:15,940] INFO The digest value is empty in snapshot (org.apache.zookeeper.server.DataTree) -[2023-11-03 19:45:15,971] INFO ZooKeeper audit is disabled. (org.apache.zookeeper.audit.ZKAuditProvider) -[2023-11-03 19:45:15,971] INFO 140 txns loaded in 26 ms (org.apache.zookeeper.server.persistence.FileTxnSnapLog) -[2023-11-03 19:45:15,971] INFO Snapshot loaded in 35 ms, highest zxid is 0x8c, digest is 307986004310 (org.apache.zookeeper.server.ZKDatabase) -[2023-11-03 19:45:15,972] INFO Snapshotting: 0x8c to /tmp/zookeeper/version-2/snapshot.8c (org.apache.zookeeper.server.persistence.FileTxnSnapLog) -[2023-11-03 19:45:15,974] INFO Snapshot taken in 2 ms (org.apache.zookeeper.server.ZooKeeperServer) -[2023-11-03 19:45:15,983] INFO zookeeper.request_throttler.shutdownTimeout = 10000 ms (org.apache.zookeeper.server.RequestThrottler) -[2023-11-03 19:45:15,983] INFO PrepRequestProcessor (sid:0) started, reconfigEnabled=false (org.apache.zookeeper.server.PrepRequestProcessor) -[2023-11-03 19:45:16,000] INFO Using checkIntervalMs=60000 maxPerMinute=10000 maxNeverUsedIntervalMs=0 (org.apache.zookeeper.server.ContainerManager) -[2023-11-03 19:45:28,280] INFO Registered kafka:type=kafka.Log4jController MBean (kafka.utils.Log4jControllerRegistration$) -[2023-11-03 19:45:28,451] INFO Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation (org.apache.zookeeper.common.X509Util) -[2023-11-03 19:45:28,513] INFO Registered signal handlers for TERM, INT, HUP (org.apache.kafka.common.utils.LoggingSignalHandler) -[2023-11-03 19:45:28,515] INFO starting (kafka.server.KafkaServer) -[2023-11-03 19:45:28,515] INFO Connecting to zookeeper on localhost:2181 (kafka.server.KafkaServer) -[2023-11-03 19:45:28,523] INFO [ZooKeeperClient Kafka server] Initializing a new session to localhost:2181. (kafka.zookeeper.ZooKeeperClient) -[2023-11-03 19:45:28,528] INFO Client environment:zookeeper.version=3.8.2-139d619b58292d7734b4fc83a0f44be4e7b0c986, built on 2023-07-05 19:24 UTC (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,529] INFO Client environment:host.name=ThinkPadP53 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,529] INFO Client environment:java.version=17.0.6 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,529] INFO Client environment:java.vendor=Eclipse Adoptium (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,529] INFO Client environment:java.home=/opt/openjdk-bin-17.0.6_p10 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,529] INFO Client environment:java.class.path=/scratch/Exemple_Kafka/bin/../libs/activation-1.1.1.jar:/scratch/Exemple_Kafka/bin/../libs/aopalliance-repackaged-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/argparse4j-0.7.0.jar:/scratch/Exemple_Kafka/bin/../libs/audience-annotations-0.12.0.jar:/scratch/Exemple_Kafka/bin/../libs/caffeine-2.9.3.jar:/scratch/Exemple_Kafka/bin/../libs/checker-qual-3.19.0.jar:/scratch/Exemple_Kafka/bin/../libs/commons-beanutils-1.9.4.jar:/scratch/Exemple_Kafka/bin/../libs/commons-cli-1.4.jar:/scratch/Exemple_Kafka/bin/../libs/commons-collections-3.2.2.jar:/scratch/Exemple_Kafka/bin/../libs/commons-digester-2.1.jar:/scratch/Exemple_Kafka/bin/../libs/commons-io-2.11.0.jar:/scratch/Exemple_Kafka/bin/../libs/commons-lang3-3.8.1.jar:/scratch/Exemple_Kafka/bin/../libs/commons-logging-1.2.jar:/scratch/Exemple_Kafka/bin/../libs/commons-validator-1.7.jar:/scratch/Exemple_Kafka/bin/../libs/connect-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-basic-auth-extension-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-json-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-mirror-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-mirror-client-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-runtime-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/connect-transforms-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/error_prone_annotations-2.10.0.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-api-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-locator-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/hk2-utils-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-annotations-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-core-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-databind-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-dataformat-csv-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-datatype-jdk8-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-jaxrs-base-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-jaxrs-json-provider-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-module-jaxb-annotations-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jackson-module-scala_2.13-2.13.5.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.activation-api-1.2.2.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.annotation-api-1.3.5.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.inject-2.6.1.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.validation-api-2.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.ws.rs-api-2.1.6.jar:/scratch/Exemple_Kafka/bin/../libs/jakarta.xml.bind-api-2.3.3.jar:/scratch/Exemple_Kafka/bin/../libs/javassist-3.29.2-GA.jar:/scratch/Exemple_Kafka/bin/../libs/javax.activation-api-1.2.0.jar:/scratch/Exemple_Kafka/bin/../libs/javax.annotation-api-1.3.2.jar:/scratch/Exemple_Kafka/bin/../libs/javax.servlet-api-3.1.0.jar:/scratch/Exemple_Kafka/bin/../libs/javax.ws.rs-api-2.1.1.jar:/scratch/Exemple_Kafka/bin/../libs/jaxb-api-2.3.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-client-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-common-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-container-servlet-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-container-servlet-core-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-hk2-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jersey-server-2.39.1.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-client-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-continuation-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-http-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-io-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-security-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-server-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-servlet-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-servlets-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-util-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jetty-util-ajax-9.4.52.v20230823.jar:/scratch/Exemple_Kafka/bin/../libs/jline-3.22.0.jar:/scratch/Exemple_Kafka/bin/../libs/jopt-simple-5.0.4.jar:/scratch/Exemple_Kafka/bin/../libs/jose4j-0.9.3.jar:/scratch/Exemple_Kafka/bin/../libs/jsr305-3.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-clients-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-group-coordinator-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-log4j-appender-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-metadata-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-raft-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-server-common-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-shell-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-storage-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-storage-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-examples-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-scala_2.13-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-streams-test-utils-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-tools-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka-tools-api-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/kafka_2.13-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/lz4-java-1.8.0.jar:/scratch/Exemple_Kafka/bin/../libs/maven-artifact-3.8.8.jar:/scratch/Exemple_Kafka/bin/../libs/metrics-core-2.2.0.jar:/scratch/Exemple_Kafka/bin/../libs/metrics-core-4.1.12.1.jar:/scratch/Exemple_Kafka/bin/../libs/netty-buffer-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-codec-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-common-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-handler-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-resolver-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-classes-epoll-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-native-epoll-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/netty-transport-native-unix-common-4.1.94.Final.jar:/scratch/Exemple_Kafka/bin/../libs/osgi-resource-locator-1.0.3.jar:/scratch/Exemple_Kafka/bin/../libs/paranamer-2.8.jar:/scratch/Exemple_Kafka/bin/../libs/pcollections-4.0.1.jar:/scratch/Exemple_Kafka/bin/../libs/plexus-utils-3.3.1.jar:/scratch/Exemple_Kafka/bin/../libs/reflections-0.10.2.jar:/scratch/Exemple_Kafka/bin/../libs/reload4j-1.2.25.jar:/scratch/Exemple_Kafka/bin/../libs/rocksdbjni-7.9.2.jar:/scratch/Exemple_Kafka/bin/../libs/scala-collection-compat_2.13-2.10.0.jar:/scratch/Exemple_Kafka/bin/../libs/scala-java8-compat_2.13-1.0.2.jar:/scratch/Exemple_Kafka/bin/../libs/scala-library-2.13.11.jar:/scratch/Exemple_Kafka/bin/../libs/scala-logging_2.13-3.9.4.jar:/scratch/Exemple_Kafka/bin/../libs/scala-reflect-2.13.11.jar:/scratch/Exemple_Kafka/bin/../libs/slf4j-api-1.7.36.jar:/scratch/Exemple_Kafka/bin/../libs/slf4j-reload4j-1.7.36.jar:/scratch/Exemple_Kafka/bin/../libs/snappy-java-1.1.10.4.jar:/scratch/Exemple_Kafka/bin/../libs/swagger-annotations-2.2.8.jar:/scratch/Exemple_Kafka/bin/../libs/trogdor-3.6.0.jar:/scratch/Exemple_Kafka/bin/../libs/zookeeper-3.8.2.jar:/scratch/Exemple_Kafka/bin/../libs/zookeeper-jute-3.8.2.jar:/scratch/Exemple_Kafka/bin/../libs/zstd-jni-1.5.5-1.jar (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,529] INFO Client environment:java.library.path=/usr/java/packages/lib:/usr/lib64:/lib64:/lib:/usr/lib (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,529] INFO Client environment:java.io.tmpdir=/tmp (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,529] INFO Client environment:java.compiler= (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,529] INFO Client environment:os.name=Linux (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,530] INFO Client environment:os.arch=amd64 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,530] INFO Client environment:os.version=6.4.3-cachyosGentooThinkPadP53 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,530] INFO Client environment:user.name=memartel (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,530] INFO Client environment:user.home=/home/memartel (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,530] INFO Client environment:user.dir=/scratch/Exemple_Kafka (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,530] INFO Client environment:os.memory.free=987MB (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,530] INFO Client environment:os.memory.max=1024MB (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,530] INFO Client environment:os.memory.total=1024MB (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,532] INFO Initiating client connection, connectString=localhost:2181 sessionTimeout=18000 watcher=kafka.zookeeper.ZooKeeperClient$ZooKeeperClientWatcher$@3fce8fd9 (org.apache.zookeeper.ZooKeeper) -[2023-11-03 19:45:28,535] INFO jute.maxbuffer value is 4194304 Bytes (org.apache.zookeeper.ClientCnxnSocket) -[2023-11-03 19:45:28,540] INFO zookeeper.request.timeout value is 0. feature enabled=false (org.apache.zookeeper.ClientCnxn) -[2023-11-03 19:45:28,542] INFO [ZooKeeperClient Kafka server] Waiting until connected. (kafka.zookeeper.ZooKeeperClient) -[2023-11-03 19:45:28,542] INFO Opening socket connection to server localhost/[0:0:0:0:0:0:0:1]:2181. (org.apache.zookeeper.ClientCnxn) -[2023-11-03 19:45:28,544] INFO Socket connection established, initiating session, client: /[0:0:0:0:0:0:0:1]:60254, server: localhost/[0:0:0:0:0:0:0:1]:2181 (org.apache.zookeeper.ClientCnxn) -[2023-11-03 19:45:28,550] INFO Creating new log file: log.8d (org.apache.zookeeper.server.persistence.FileTxnLog) -[2023-11-03 19:45:28,553] INFO Session establishment complete on server localhost/[0:0:0:0:0:0:0:1]:2181, session id = 0x100008b139a0000, negotiated timeout = 18000 (org.apache.zookeeper.ClientCnxn) -[2023-11-03 19:45:28,555] INFO [ZooKeeperClient Kafka server] Connected. (kafka.zookeeper.ZooKeeperClient) -[2023-11-03 19:45:28,704] INFO Cluster ID = 8EdUX-SkTwmB2aSLdVSXoQ (kafka.server.KafkaServer) -[2023-11-03 19:45:28,753] INFO KafkaConfig values: - advertised.listeners = null - alter.config.policy.class.name = null - alter.log.dirs.replication.quota.window.num = 11 - alter.log.dirs.replication.quota.window.size.seconds = 1 - authorizer.class.name = - auto.create.topics.enable = true - auto.include.jmx.reporter = true - auto.leader.rebalance.enable = true - background.threads = 10 - broker.heartbeat.interval.ms = 2000 - broker.id = 0 - broker.id.generation.enable = true - broker.rack = null - broker.session.timeout.ms = 9000 - client.quota.callback.class = null - compression.type = producer - connection.failed.authentication.delay.ms = 100 - connections.max.idle.ms = 600000 - connections.max.reauth.ms = 0 - control.plane.listener.name = null - controlled.shutdown.enable = true - controlled.shutdown.max.retries = 3 - controlled.shutdown.retry.backoff.ms = 5000 - controller.listener.names = null - controller.quorum.append.linger.ms = 25 - controller.quorum.election.backoff.max.ms = 1000 - controller.quorum.election.timeout.ms = 1000 - controller.quorum.fetch.timeout.ms = 2000 - controller.quorum.request.timeout.ms = 2000 - controller.quorum.retry.backoff.ms = 20 - controller.quorum.voters = [] - controller.quota.window.num = 11 - controller.quota.window.size.seconds = 1 - controller.socket.timeout.ms = 30000 - create.topic.policy.class.name = null - default.replication.factor = 1 - delegation.token.expiry.check.interval.ms = 3600000 - delegation.token.expiry.time.ms = 86400000 - delegation.token.master.key = null - delegation.token.max.lifetime.ms = 604800000 - delegation.token.secret.key = null - delete.records.purgatory.purge.interval.requests = 1 - delete.topic.enable = true - early.start.listeners = null - fetch.max.bytes = 57671680 - fetch.purgatory.purge.interval.requests = 1000 - group.consumer.assignors = [org.apache.kafka.coordinator.group.assignor.RangeAssignor] - group.consumer.heartbeat.interval.ms = 5000 - group.consumer.max.heartbeat.interval.ms = 15000 - group.consumer.max.session.timeout.ms = 60000 - group.consumer.max.size = 2147483647 - group.consumer.min.heartbeat.interval.ms = 5000 - group.consumer.min.session.timeout.ms = 45000 - group.consumer.session.timeout.ms = 45000 - group.coordinator.new.enable = false - group.coordinator.threads = 1 - group.initial.rebalance.delay.ms = 0 - group.max.session.timeout.ms = 1800000 - group.max.size = 2147483647 - group.min.session.timeout.ms = 6000 - initial.broker.registration.timeout.ms = 60000 - inter.broker.listener.name = null - inter.broker.protocol.version = 3.6-IV2 - kafka.metrics.polling.interval.secs = 10 - kafka.metrics.reporters = [] - leader.imbalance.check.interval.seconds = 300 - leader.imbalance.per.broker.percentage = 10 - listener.security.protocol.map = PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL - listeners = PLAINTEXT://:9092 - log.cleaner.backoff.ms = 15000 - log.cleaner.dedupe.buffer.size = 134217728 - log.cleaner.delete.retention.ms = 86400000 - log.cleaner.enable = true - log.cleaner.io.buffer.load.factor = 0.9 - log.cleaner.io.buffer.size = 524288 - log.cleaner.io.max.bytes.per.second = 1.7976931348623157E308 - log.cleaner.max.compaction.lag.ms = 9223372036854775807 - log.cleaner.min.cleanable.ratio = 0.5 - log.cleaner.min.compaction.lag.ms = 0 - log.cleaner.threads = 1 - log.cleanup.policy = [delete] - log.dir = /tmp/kafka-logs - log.dirs = /tmp/kafka-logs - log.flush.interval.messages = 9223372036854775807 - log.flush.interval.ms = null - log.flush.offset.checkpoint.interval.ms = 60000 - log.flush.scheduler.interval.ms = 9223372036854775807 - log.flush.start.offset.checkpoint.interval.ms = 60000 - log.index.interval.bytes = 4096 - log.index.size.max.bytes = 10485760 - log.local.retention.bytes = -2 - log.local.retention.ms = -2 - log.message.downconversion.enable = true - log.message.format.version = 3.0-IV1 - log.message.timestamp.after.max.ms = 9223372036854775807 - log.message.timestamp.before.max.ms = 9223372036854775807 - log.message.timestamp.difference.max.ms = 9223372036854775807 - log.message.timestamp.type = CreateTime - log.preallocate = false - log.retention.bytes = -1 - log.retention.check.interval.ms = 300000 - log.retention.hours = 168 - log.retention.minutes = null - log.retention.ms = null - log.roll.hours = 168 - log.roll.jitter.hours = 0 - log.roll.jitter.ms = null - log.roll.ms = null - log.segment.bytes = 1073741824 - log.segment.delete.delay.ms = 60000 - max.connection.creation.rate = 2147483647 - max.connections = 2147483647 - max.connections.per.ip = 2147483647 - max.connections.per.ip.overrides = - max.incremental.fetch.session.cache.slots = 1000 - message.max.bytes = 1048588 - metadata.log.dir = null - metadata.log.max.record.bytes.between.snapshots = 20971520 - metadata.log.max.snapshot.interval.ms = 3600000 - metadata.log.segment.bytes = 1073741824 - metadata.log.segment.min.bytes = 8388608 - metadata.log.segment.ms = 604800000 - metadata.max.idle.interval.ms = 500 - metadata.max.retention.bytes = 104857600 - metadata.max.retention.ms = 604800000 - metric.reporters = [] - metrics.num.samples = 2 - metrics.recording.level = INFO - metrics.sample.window.ms = 30000 - min.insync.replicas = 1 - node.id = 0 - num.io.threads = 8 - num.network.threads = 3 - num.partitions = 1 - num.recovery.threads.per.data.dir = 1 - num.replica.alter.log.dirs.threads = null - num.replica.fetchers = 1 - offset.metadata.max.bytes = 4096 - offsets.commit.required.acks = -1 - offsets.commit.timeout.ms = 5000 - offsets.load.buffer.size = 5242880 - offsets.retention.check.interval.ms = 600000 - offsets.retention.minutes = 10080 - offsets.topic.compression.codec = 0 - offsets.topic.num.partitions = 50 - offsets.topic.replication.factor = 1 - offsets.topic.segment.bytes = 104857600 - password.encoder.cipher.algorithm = AES/CBC/PKCS5Padding - password.encoder.iterations = 4096 - password.encoder.key.length = 128 - password.encoder.keyfactory.algorithm = null - password.encoder.old.secret = null - password.encoder.secret = null - principal.builder.class = class org.apache.kafka.common.security.authenticator.DefaultKafkaPrincipalBuilder - process.roles = [] - producer.id.expiration.check.interval.ms = 600000 - producer.id.expiration.ms = 86400000 - producer.purgatory.purge.interval.requests = 1000 - queued.max.request.bytes = -1 - queued.max.requests = 500 - quota.window.num = 11 - quota.window.size.seconds = 1 - remote.log.index.file.cache.total.size.bytes = 1073741824 - remote.log.manager.task.interval.ms = 30000 - remote.log.manager.task.retry.backoff.max.ms = 30000 - remote.log.manager.task.retry.backoff.ms = 500 - remote.log.manager.task.retry.jitter = 0.2 - remote.log.manager.thread.pool.size = 10 - remote.log.metadata.custom.metadata.max.bytes = 128 - remote.log.metadata.manager.class.name = org.apache.kafka.server.log.remote.metadata.storage.TopicBasedRemoteLogMetadataManager - remote.log.metadata.manager.class.path = null - remote.log.metadata.manager.impl.prefix = rlmm.config. - remote.log.metadata.manager.listener.name = null - remote.log.reader.max.pending.tasks = 100 - remote.log.reader.threads = 10 - remote.log.storage.manager.class.name = null - remote.log.storage.manager.class.path = null - remote.log.storage.manager.impl.prefix = rsm.config. - remote.log.storage.system.enable = false - replica.fetch.backoff.ms = 1000 - replica.fetch.max.bytes = 1048576 - replica.fetch.min.bytes = 1 - replica.fetch.response.max.bytes = 10485760 - replica.fetch.wait.max.ms = 500 - replica.high.watermark.checkpoint.interval.ms = 5000 - replica.lag.time.max.ms = 30000 - replica.selector.class = null - replica.socket.receive.buffer.bytes = 65536 - replica.socket.timeout.ms = 30000 - replication.quota.window.num = 11 - replication.quota.window.size.seconds = 1 - request.timeout.ms = 30000 - reserved.broker.max.id = 1000 - sasl.client.callback.handler.class = null - sasl.enabled.mechanisms = [GSSAPI] - sasl.jaas.config = null - sasl.kerberos.kinit.cmd = /usr/bin/kinit - sasl.kerberos.min.time.before.relogin = 60000 - sasl.kerberos.principal.to.local.rules = [DEFAULT] - sasl.kerberos.service.name = null - sasl.kerberos.ticket.renew.jitter = 0.05 - sasl.kerberos.ticket.renew.window.factor = 0.8 - sasl.login.callback.handler.class = null - sasl.login.class = null - sasl.login.connect.timeout.ms = null - sasl.login.read.timeout.ms = null - sasl.login.refresh.buffer.seconds = 300 - sasl.login.refresh.min.period.seconds = 60 - sasl.login.refresh.window.factor = 0.8 - sasl.login.refresh.window.jitter = 0.05 - sasl.login.retry.backoff.max.ms = 10000 - sasl.login.retry.backoff.ms = 100 - sasl.mechanism.controller.protocol = GSSAPI - sasl.mechanism.inter.broker.protocol = GSSAPI - sasl.oauthbearer.clock.skew.seconds = 30 - sasl.oauthbearer.expected.audience = null - sasl.oauthbearer.expected.issuer = null - sasl.oauthbearer.jwks.endpoint.refresh.ms = 3600000 - sasl.oauthbearer.jwks.endpoint.retry.backoff.max.ms = 10000 - sasl.oauthbearer.jwks.endpoint.retry.backoff.ms = 100 - sasl.oauthbearer.jwks.endpoint.url = null - sasl.oauthbearer.scope.claim.name = scope - sasl.oauthbearer.sub.claim.name = sub - sasl.oauthbearer.token.endpoint.url = null - sasl.server.callback.handler.class = null - sasl.server.max.receive.size = 524288 - security.inter.broker.protocol = PLAINTEXT - security.providers = null - server.max.startup.time.ms = 9223372036854775807 - socket.connection.setup.timeout.max.ms = 30000 - socket.connection.setup.timeout.ms = 10000 - socket.listen.backlog.size = 50 - socket.receive.buffer.bytes = 102400 - socket.request.max.bytes = 104857600 - socket.send.buffer.bytes = 102400 - ssl.cipher.suites = [] - ssl.client.auth = none - ssl.enabled.protocols = [TLSv1.2, TLSv1.3] - ssl.endpoint.identification.algorithm = https - ssl.engine.factory.class = null - ssl.key.password = null - ssl.keymanager.algorithm = SunX509 - ssl.keystore.certificate.chain = null - ssl.keystore.key = null - ssl.keystore.location = null - ssl.keystore.password = null - ssl.keystore.type = JKS - ssl.principal.mapping.rules = DEFAULT - ssl.protocol = TLSv1.3 - ssl.provider = null - ssl.secure.random.implementation = null - ssl.trustmanager.algorithm = PKIX - ssl.truststore.certificates = null - ssl.truststore.location = null - ssl.truststore.password = null - ssl.truststore.type = JKS - transaction.abort.timed.out.transaction.cleanup.interval.ms = 10000 - transaction.max.timeout.ms = 900000 - transaction.partition.verification.enable = true - transaction.remove.expired.transaction.cleanup.interval.ms = 3600000 - transaction.state.log.load.buffer.size = 5242880 - transaction.state.log.min.isr = 1 - transaction.state.log.num.partitions = 50 - transaction.state.log.replication.factor = 1 - transaction.state.log.segment.bytes = 104857600 - transactional.id.expiration.ms = 604800000 - unclean.leader.election.enable = false - unstable.api.versions.enable = false - zookeeper.clientCnxnSocket = null - zookeeper.connect = localhost:2181 - zookeeper.connection.timeout.ms = 18000 - zookeeper.max.in.flight.requests = 10 - zookeeper.metadata.migration.enable = false - zookeeper.session.timeout.ms = 18000 - zookeeper.set.acl = false - zookeeper.ssl.cipher.suites = null - zookeeper.ssl.client.enable = false - zookeeper.ssl.crl.enable = false - zookeeper.ssl.enabled.protocols = null - zookeeper.ssl.endpoint.identification.algorithm = HTTPS - zookeeper.ssl.keystore.location = null - zookeeper.ssl.keystore.password = null - zookeeper.ssl.keystore.type = null - zookeeper.ssl.ocsp.enable = false - zookeeper.ssl.protocol = TLSv1.2 - zookeeper.ssl.truststore.location = null - zookeeper.ssl.truststore.password = null - zookeeper.ssl.truststore.type = null - (kafka.server.KafkaConfig) -[2023-11-03 19:45:28,782] INFO [ThrottledChannelReaper-Fetch]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:45:28,782] INFO [ThrottledChannelReaper-Produce]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:45:28,783] INFO [ThrottledChannelReaper-Request]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:45:28,784] INFO [ThrottledChannelReaper-ControllerMutation]: Starting (kafka.server.ClientQuotaManager$ThrottledChannelReaper) -[2023-11-03 19:45:28,816] INFO Loading logs from log dirs ArraySeq(/tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,825] INFO Skipping recovery of 51 logs from /tmp/kafka-logs since clean shutdown file was found (kafka.log.LogManager) -[2023-11-03 19:45:28,862] INFO [LogLoader partition=test-topic-0, dir=/tmp/kafka-logs] Loading producer state till offset 5 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,870] INFO [LogLoader partition=test-topic-0, dir=/tmp/kafka-logs] Reloading from producer snapshot and rebuilding producer state from offset 5 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,870] INFO [ProducerStateManager partition=test-topic-0]Loading producer state from snapshot file 'SnapshotFile(offset=5, file=/tmp/kafka-logs/test-topic-0/00000000000000000005.snapshot)' (org.apache.kafka.storage.internals.log.ProducerStateManager) -[2023-11-03 19:45:28,880] INFO [LogLoader partition=test-topic-0, dir=/tmp/kafka-logs] Producer state recovery took 10ms for snapshot load and 0ms for segment recovery from offset 5 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,894] INFO Completed load of Log(dir=/tmp/kafka-logs/test-topic-0, topicId=Pb9zfnlKRkmTGaMQyAABkw, topic=test-topic, partition=0, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=5) with 1 segments, local-log-start-offset 0 and log-end-offset 5 in 61ms (1/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,897] INFO [LogLoader partition=__consumer_offsets-28, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,899] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-28, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=28, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 5ms (2/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,901] INFO [LogLoader partition=__consumer_offsets-13, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,903] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-13, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=13, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (3/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,905] INFO [LogLoader partition=__consumer_offsets-43, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,907] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-43, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=43, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 4ms (4/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,911] INFO [LogLoader partition=__consumer_offsets-6, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,913] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-6, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=6, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 6ms (5/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,915] INFO [LogLoader partition=__consumer_offsets-36, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,917] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-36, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=36, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 4ms (6/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,919] INFO [LogLoader partition=__consumer_offsets-21, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,920] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-21, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=21, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (7/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,922] INFO [LogLoader partition=__consumer_offsets-12, dir=/tmp/kafka-logs] Loading producer state till offset 7 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,922] INFO [LogLoader partition=__consumer_offsets-12, dir=/tmp/kafka-logs] Reloading from producer snapshot and rebuilding producer state from offset 7 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,922] INFO [ProducerStateManager partition=__consumer_offsets-12]Loading producer state from snapshot file 'SnapshotFile(offset=7, file=/tmp/kafka-logs/__consumer_offsets-12/00000000000000000007.snapshot)' (org.apache.kafka.storage.internals.log.ProducerStateManager) -[2023-11-03 19:45:28,923] INFO [LogLoader partition=__consumer_offsets-12, dir=/tmp/kafka-logs] Producer state recovery took 1ms for snapshot load and 0ms for segment recovery from offset 7 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,924] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-12, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=12, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=7) with 1 segments, local-log-start-offset 0 and log-end-offset 7 in 4ms (8/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,927] INFO [LogLoader partition=__consumer_offsets-42, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,928] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-42, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=42, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (9/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,930] INFO [LogLoader partition=__consumer_offsets-27, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,932] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-27, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=27, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (10/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,934] INFO [LogLoader partition=__consumer_offsets-20, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,935] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-20, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=20, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (11/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,937] INFO [LogLoader partition=__consumer_offsets-5, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,938] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-5, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=5, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (12/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,940] INFO [LogLoader partition=__consumer_offsets-35, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,942] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-35, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=35, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (13/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,944] INFO [LogLoader partition=__consumer_offsets-0, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,946] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-0, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=0, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 4ms (14/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,948] INFO [LogLoader partition=__consumer_offsets-30, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,949] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-30, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=30, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (15/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,951] INFO [LogLoader partition=__consumer_offsets-15, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,952] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-15, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=15, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (16/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,954] INFO [LogLoader partition=__consumer_offsets-45, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,955] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-45, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=45, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (17/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,958] INFO [LogLoader partition=__consumer_offsets-8, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,960] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-8, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=8, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 5ms (18/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,962] INFO [LogLoader partition=__consumer_offsets-38, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,963] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-38, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=38, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (19/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,965] INFO [LogLoader partition=__consumer_offsets-23, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,966] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-23, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=23, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (20/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,968] INFO [LogLoader partition=__consumer_offsets-14, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,969] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-14, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=14, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (21/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,970] INFO [LogLoader partition=__consumer_offsets-44, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,971] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-44, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=44, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (22/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,973] INFO [LogLoader partition=__consumer_offsets-29, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,975] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-29, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=29, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 4ms (23/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,978] INFO [LogLoader partition=__consumer_offsets-22, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,980] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-22, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=22, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 5ms (24/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,981] INFO [LogLoader partition=__consumer_offsets-7, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,982] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-7, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=7, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (25/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,984] INFO [LogLoader partition=__consumer_offsets-37, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,985] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-37, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=37, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (26/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,987] INFO [LogLoader partition=__consumer_offsets-32, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,987] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-32, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=32, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (27/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,989] INFO [LogLoader partition=__consumer_offsets-17, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,991] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-17, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=17, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (28/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,995] INFO [LogLoader partition=__consumer_offsets-47, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,996] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-47, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=47, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 4ms (29/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:28,998] INFO [LogLoader partition=__consumer_offsets-40, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:28,999] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-40, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=40, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (30/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,001] INFO [LogLoader partition=__consumer_offsets-25, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,002] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-25, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=25, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (31/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,004] INFO [LogLoader partition=__consumer_offsets-2, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,005] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-2, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=2, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (32/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,006] INFO [LogLoader partition=__consumer_offsets-16, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,008] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-16, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=16, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (33/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,011] INFO [LogLoader partition=__consumer_offsets-1, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,012] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-1, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=1, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 5ms (34/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,014] INFO [LogLoader partition=__consumer_offsets-46, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,016] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-46, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=46, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (35/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,017] INFO [LogLoader partition=__consumer_offsets-31, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,018] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-31, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=31, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (36/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,019] INFO [LogLoader partition=__consumer_offsets-24, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,020] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-24, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=24, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (37/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,022] INFO [LogLoader partition=__consumer_offsets-9, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,023] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-9, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=9, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (38/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,026] INFO [LogLoader partition=__consumer_offsets-39, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,027] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-39, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=39, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 4ms (39/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,029] INFO [LogLoader partition=__consumer_offsets-49, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,030] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-49, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=49, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (40/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,032] INFO [LogLoader partition=__consumer_offsets-26, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,033] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-26, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=26, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (41/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,035] INFO [LogLoader partition=__consumer_offsets-11, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,036] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-11, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=11, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 3ms (42/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,037] INFO [LogLoader partition=__consumer_offsets-4, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,038] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-4, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=4, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (43/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,040] INFO [LogLoader partition=__consumer_offsets-34, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,042] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-34, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=34, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 4ms (44/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,044] INFO [LogLoader partition=__consumer_offsets-19, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,045] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-19, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=19, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (45/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,046] INFO [LogLoader partition=__consumer_offsets-48, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,046] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-48, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=48, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (46/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,047] INFO [LogLoader partition=__consumer_offsets-33, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,048] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-33, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=33, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 1ms (47/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,049] INFO [LogLoader partition=__consumer_offsets-10, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,050] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-10, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=10, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (48/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,051] INFO [LogLoader partition=__consumer_offsets-41, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,052] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-41, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=41, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (49/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,053] INFO [LogLoader partition=__consumer_offsets-18, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,053] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-18, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=18, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 2ms (50/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,054] INFO [LogLoader partition=__consumer_offsets-3, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) -[2023-11-03 19:45:29,055] INFO Completed load of Log(dir=/tmp/kafka-logs/__consumer_offsets-3, topicId=vfIIZeo7TYKNeTmKGzMlyg, topic=__consumer_offsets, partition=3, highWatermark=0, lastStableOffset=0, logStartOffset=0, logEndOffset=0) with 1 segments, local-log-start-offset 0 and log-end-offset 0 in 1ms (51/51 completed in /tmp/kafka-logs) (kafka.log.LogManager) -[2023-11-03 19:45:29,058] INFO Loaded 51 logs in 241ms (kafka.log.LogManager) -[2023-11-03 19:45:29,060] INFO Starting log cleanup with a period of 300000 ms. (kafka.log.LogManager) -[2023-11-03 19:45:29,060] INFO Starting log flusher with a default period of 9223372036854775807 ms. (kafka.log.LogManager) -[2023-11-03 19:45:29,089] INFO [kafka-log-cleaner-thread-0]: Starting (kafka.log.LogCleaner$CleanerThread) -[2023-11-03 19:45:29,102] INFO [feature-zk-node-event-process-thread]: Starting (kafka.server.FinalizedFeatureChangeListener$ChangeNotificationProcessorThread) -[2023-11-03 19:45:29,115] INFO [MetadataCache brokerId=0] Updated cache from existing None to latest Features(version=3.6-IV2, finalizedFeatures={}, finalizedFeaturesEpoch=0). (kafka.server.metadata.ZkMetadataCache) -[2023-11-03 19:45:29,131] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Starting (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:45:29,305] INFO Updated connection-accept-rate max connection creation rate to 2147483647 (kafka.network.ConnectionQuotas) -[2023-11-03 19:45:29,317] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Created data-plane acceptor and processors for endpoint : ListenerName(PLAINTEXT) (kafka.network.SocketServer) -[2023-11-03 19:45:29,319] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Starting (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:45:29,335] INFO [ExpirationReaper-0-Produce]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:45:29,336] INFO [ExpirationReaper-0-Fetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:45:29,336] INFO [ExpirationReaper-0-DeleteRecords]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:45:29,336] INFO [ExpirationReaper-0-ElectLeader]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:45:29,337] INFO [ExpirationReaper-0-RemoteFetch]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:45:29,346] INFO [LogDirFailureHandler]: Starting (kafka.server.ReplicaManager$LogDirFailureHandler) -[2023-11-03 19:45:29,346] INFO [AddPartitionsToTxnSenderThread-0]: Starting (kafka.server.AddPartitionsToTxnManager) -[2023-11-03 19:45:29,373] INFO Creating /brokers/ids/0 (is it secure? false) (kafka.zk.KafkaZkClient) -[2023-11-03 19:45:29,390] INFO Stat of the created znode at /brokers/ids/0 is: 156,156,1699055129385,1699055129385,1,0,0,72058191367241728,206,0,156 - (kafka.zk.KafkaZkClient) -[2023-11-03 19:45:29,390] INFO Registered broker 0 at path /brokers/ids/0 with addresses: PLAINTEXT://ThinkPadP53:9092, czxid (broker epoch): 156 (kafka.zk.KafkaZkClient) -[2023-11-03 19:45:29,432] INFO [ExpirationReaper-0-topic]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:45:29,436] INFO [ExpirationReaper-0-Heartbeat]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:45:29,437] INFO [ExpirationReaper-0-Rebalance]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:45:29,451] INFO [GroupCoordinator 0]: Starting up. (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,462] INFO [GroupCoordinator 0]: Startup complete. (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,473] INFO [TransactionCoordinator id=0] Starting up. (kafka.coordinator.transaction.TransactionCoordinator) -[2023-11-03 19:45:29,479] INFO [TxnMarkerSenderThread-0]: Starting (kafka.coordinator.transaction.TransactionMarkerChannelManager) -[2023-11-03 19:45:29,479] INFO [TransactionCoordinator id=0] Startup complete. (kafka.coordinator.transaction.TransactionCoordinator) -[2023-11-03 19:45:29,515] INFO [ExpirationReaper-0-AlterAcls]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) -[2023-11-03 19:45:29,536] INFO [Controller id=0, targetBrokerId=0] Node 0 disconnected. (org.apache.kafka.clients.NetworkClient) -[2023-11-03 19:45:29,538] INFO [/config/changes-event-process-thread]: Starting (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread) -[2023-11-03 19:45:29,538] WARN [Controller id=0, targetBrokerId=0] Connection to node 0 (ThinkPadP53/192.168.37.3:9092) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient) -[2023-11-03 19:45:29,541] INFO [Controller id=0, targetBrokerId=0] Client requested connection close from node 0 (org.apache.kafka.clients.NetworkClient) -[2023-11-03 19:45:29,559] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Enabling request processing. (kafka.network.SocketServer) -[2023-11-03 19:45:29,562] INFO Awaiting socket connections on 0.0.0.0:9092. (kafka.network.DataPlaneAcceptor) -[2023-11-03 19:45:29,567] INFO Kafka version: 3.6.0 (org.apache.kafka.common.utils.AppInfoParser) -[2023-11-03 19:45:29,567] INFO Kafka commitId: 60e845626d8a465a (org.apache.kafka.common.utils.AppInfoParser) -[2023-11-03 19:45:29,567] INFO Kafka startTimeMs: 1699055129564 (org.apache.kafka.common.utils.AppInfoParser) -[2023-11-03 19:45:29,568] INFO [KafkaServer id=0] started (kafka.server.KafkaServer) -[2023-11-03 19:45:29,710] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions HashSet(__consumer_offsets-22, __consumer_offsets-30, __consumer_offsets-25, __consumer_offsets-35, __consumer_offsets-37, __consumer_offsets-38, __consumer_offsets-13, test-topic-0, __consumer_offsets-8, __consumer_offsets-21, __consumer_offsets-4, __consumer_offsets-27, __consumer_offsets-7, __consumer_offsets-9, __consumer_offsets-46, __consumer_offsets-41, __consumer_offsets-33, __consumer_offsets-23, __consumer_offsets-49, __consumer_offsets-47, __consumer_offsets-16, __consumer_offsets-28, __consumer_offsets-31, __consumer_offsets-36, __consumer_offsets-42, __consumer_offsets-3, __consumer_offsets-18, __consumer_offsets-15, __consumer_offsets-24, __consumer_offsets-17, __consumer_offsets-48, __consumer_offsets-19, __consumer_offsets-11, __consumer_offsets-2, __consumer_offsets-43, __consumer_offsets-6, __consumer_offsets-14, __consumer_offsets-20, __consumer_offsets-0, __consumer_offsets-44, __consumer_offsets-39, __consumer_offsets-12, __consumer_offsets-45, __consumer_offsets-1, __consumer_offsets-5, __consumer_offsets-26, __consumer_offsets-29, __consumer_offsets-34, __consumer_offsets-10, __consumer_offsets-32, __consumer_offsets-40) (kafka.server.ReplicaFetcherManager) -[2023-11-03 19:45:29,717] INFO [Partition __consumer_offsets-3 broker=0] Log loaded for partition __consumer_offsets-3 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,721] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Recorded new controller, from now on will use node ThinkPadP53:9092 (id: 0 rack: null) (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:45:29,721] INFO [Partition __consumer_offsets-18 broker=0] Log loaded for partition __consumer_offsets-18 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,722] INFO [Partition __consumer_offsets-41 broker=0] Log loaded for partition __consumer_offsets-41 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,722] INFO [Partition __consumer_offsets-10 broker=0] Log loaded for partition __consumer_offsets-10 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,722] INFO [Partition __consumer_offsets-33 broker=0] Log loaded for partition __consumer_offsets-33 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,723] INFO [Partition __consumer_offsets-48 broker=0] Log loaded for partition __consumer_offsets-48 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,723] INFO [Partition __consumer_offsets-19 broker=0] Log loaded for partition __consumer_offsets-19 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,723] INFO [Partition __consumer_offsets-34 broker=0] Log loaded for partition __consumer_offsets-34 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,724] INFO [Partition __consumer_offsets-4 broker=0] Log loaded for partition __consumer_offsets-4 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,724] INFO [Partition __consumer_offsets-11 broker=0] Log loaded for partition __consumer_offsets-11 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,724] INFO [Partition __consumer_offsets-26 broker=0] Log loaded for partition __consumer_offsets-26 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,724] INFO [Partition __consumer_offsets-49 broker=0] Log loaded for partition __consumer_offsets-49 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,725] INFO [Partition __consumer_offsets-39 broker=0] Log loaded for partition __consumer_offsets-39 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,725] INFO [Partition __consumer_offsets-9 broker=0] Log loaded for partition __consumer_offsets-9 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,725] INFO [Partition __consumer_offsets-24 broker=0] Log loaded for partition __consumer_offsets-24 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,726] INFO [Partition __consumer_offsets-31 broker=0] Log loaded for partition __consumer_offsets-31 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,726] INFO [Partition __consumer_offsets-46 broker=0] Log loaded for partition __consumer_offsets-46 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,726] INFO [Partition __consumer_offsets-1 broker=0] Log loaded for partition __consumer_offsets-1 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,727] INFO [Partition __consumer_offsets-16 broker=0] Log loaded for partition __consumer_offsets-16 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,727] INFO [Partition __consumer_offsets-2 broker=0] Log loaded for partition __consumer_offsets-2 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,728] INFO [Partition __consumer_offsets-25 broker=0] Log loaded for partition __consumer_offsets-25 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,728] INFO [Partition __consumer_offsets-40 broker=0] Log loaded for partition __consumer_offsets-40 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,728] INFO [Partition __consumer_offsets-47 broker=0] Log loaded for partition __consumer_offsets-47 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,729] INFO [Partition __consumer_offsets-17 broker=0] Log loaded for partition __consumer_offsets-17 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,729] INFO [Partition __consumer_offsets-32 broker=0] Log loaded for partition __consumer_offsets-32 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,730] INFO [Partition __consumer_offsets-37 broker=0] Log loaded for partition __consumer_offsets-37 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,730] INFO [Partition __consumer_offsets-7 broker=0] Log loaded for partition __consumer_offsets-7 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,730] INFO [Partition __consumer_offsets-22 broker=0] Log loaded for partition __consumer_offsets-22 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,730] INFO [Partition __consumer_offsets-29 broker=0] Log loaded for partition __consumer_offsets-29 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,731] INFO [Partition __consumer_offsets-44 broker=0] Log loaded for partition __consumer_offsets-44 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,731] INFO [Partition __consumer_offsets-14 broker=0] Log loaded for partition __consumer_offsets-14 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,731] INFO [Partition __consumer_offsets-23 broker=0] Log loaded for partition __consumer_offsets-23 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,732] INFO [Partition __consumer_offsets-38 broker=0] Log loaded for partition __consumer_offsets-38 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,732] INFO [Partition __consumer_offsets-8 broker=0] Log loaded for partition __consumer_offsets-8 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,732] INFO [Partition __consumer_offsets-45 broker=0] Log loaded for partition __consumer_offsets-45 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,732] INFO [Partition __consumer_offsets-15 broker=0] Log loaded for partition __consumer_offsets-15 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,733] INFO [Partition __consumer_offsets-30 broker=0] Log loaded for partition __consumer_offsets-30 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,733] INFO [Partition __consumer_offsets-0 broker=0] Log loaded for partition __consumer_offsets-0 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,733] INFO [Partition __consumer_offsets-35 broker=0] Log loaded for partition __consumer_offsets-35 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,734] INFO [Partition __consumer_offsets-5 broker=0] Log loaded for partition __consumer_offsets-5 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,734] INFO [Partition __consumer_offsets-20 broker=0] Log loaded for partition __consumer_offsets-20 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,734] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Recorded new controller, from now on will use node ThinkPadP53:9092 (id: 0 rack: null) (kafka.server.BrokerToControllerRequestThread) -[2023-11-03 19:45:29,734] INFO [Partition __consumer_offsets-27 broker=0] Log loaded for partition __consumer_offsets-27 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,734] INFO [Partition __consumer_offsets-42 broker=0] Log loaded for partition __consumer_offsets-42 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,734] INFO [Partition __consumer_offsets-12 broker=0] Log loaded for partition __consumer_offsets-12 with initial high watermark 7 (kafka.cluster.Partition) -[2023-11-03 19:45:29,735] INFO [Partition __consumer_offsets-21 broker=0] Log loaded for partition __consumer_offsets-21 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,735] INFO [Partition test-topic-0 broker=0] Log loaded for partition test-topic-0 with initial high watermark 5 (kafka.cluster.Partition) -[2023-11-03 19:45:29,735] INFO [Partition __consumer_offsets-36 broker=0] Log loaded for partition __consumer_offsets-36 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,736] INFO [Partition __consumer_offsets-6 broker=0] Log loaded for partition __consumer_offsets-6 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,736] INFO [Partition __consumer_offsets-43 broker=0] Log loaded for partition __consumer_offsets-43 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,736] INFO [Partition __consumer_offsets-13 broker=0] Log loaded for partition __consumer_offsets-13 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,736] INFO [Partition __consumer_offsets-28 broker=0] Log loaded for partition __consumer_offsets-28 with initial high watermark 0 (kafka.cluster.Partition) -[2023-11-03 19:45:29,740] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 3 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,741] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-3 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 18 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-18 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 41 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-41 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 10 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-10 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 33 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-33 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 48 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-48 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 19 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-19 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 34 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-34 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 4 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-4 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 11 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-11 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 26 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-26 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 49 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-49 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 39 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-39 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,743] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 9 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,743] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-9 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 24 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-24 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 31 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-31 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 46 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-46 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 1 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-1 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 16 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-16 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 2 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-2 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 25 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-25 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 40 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-40 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 47 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-47 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 17 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-17 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 32 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-32 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 37 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-37 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 7 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-7 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,744] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 22 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,744] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-22 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 29 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-29 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 44 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-44 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 14 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-14 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 23 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-23 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 38 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-38 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 8 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-8 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 45 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-45 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 15 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-15 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 30 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-30 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 0 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-0 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 35 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-35 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 5 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-5 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 20 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-20 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 27 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-27 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 42 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-42 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 12 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-12 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 21 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-21 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,745] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 36 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,745] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-36 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,746] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 6 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,746] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-6 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,746] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 43 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,746] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-43 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,746] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 13 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,746] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-13 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,746] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 28 in epoch 0 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,746] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-28 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,747] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-3 in 5 milliseconds for epoch 0, of which 2 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,747] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-18 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,748] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-41 in 5 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,748] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-10 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,748] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-33 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,748] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-48 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,748] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-19 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,748] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-34 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,748] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-4 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,749] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-11 in 6 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,749] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-26 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,749] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-49 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,749] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-39 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,749] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-9 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,749] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-24 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,749] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-31 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,749] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-46 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,750] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-1 in 6 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,750] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-16 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,750] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-2 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,750] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-25 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,750] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-40 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,750] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-47 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,750] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-17 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,750] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-32 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,751] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-37 in 7 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,751] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-7 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,751] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-22 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,751] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-29 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,751] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-44 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,751] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-14 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,751] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-23 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,751] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-38 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,751] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-8 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,752] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-45 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,752] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-15 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,752] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-30 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,752] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-0 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,752] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-35 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,752] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-5 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,752] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-20 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,752] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-27 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,752] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-42 in 7 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,771] INFO Loaded member MemberMetadata(memberId=rdkafka-d8d257cb-90fd-4aed-a191-0c8705f20f1a, groupInstanceId=None, clientId=rdkafka, clientHost=/192.168.37.3, sessionTimeoutMs=45000, rebalanceTimeoutMs=300000, supportedProtocols=List(range)) in group test-group with generation 1. (kafka.coordinator.group.GroupMetadata$) -[2023-11-03 19:45:29,781] INFO [GroupCoordinator 0]: Loading group metadata for test-group with generation 2 (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:29,782] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-12 in 37 milliseconds for epoch 0, of which 7 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,782] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-21 in 37 milliseconds for epoch 0, of which 37 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,782] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-36 in 36 milliseconds for epoch 0, of which 36 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,782] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-6 in 36 milliseconds for epoch 0, of which 36 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,782] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-43 in 36 milliseconds for epoch 0, of which 36 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,783] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-13 in 37 milliseconds for epoch 0, of which 36 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:29,783] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-28 in 37 milliseconds for epoch 0, of which 37 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) -[2023-11-03 19:45:47,781] INFO [GroupCoordinator 0]: Dynamic member with unknown member id joins group test-group in Empty state. Created a new member id rdkafka-cf033a01-fd53-46c7-9ac9-d92cbfe8c9d8 and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:47,789] INFO [GroupCoordinator 0]: Preparing to rebalance group test-group in state PreparingRebalance with old generation 2 (__consumer_offsets-12) (reason: Adding new member rdkafka-cf033a01-fd53-46c7-9ac9-d92cbfe8c9d8 with group instance id None; client reason: not provided) (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:47,795] INFO [GroupCoordinator 0]: Stabilized group test-group generation 3 (__consumer_offsets-12) with 1 members (kafka.coordinator.group.GroupCoordinator) -[2023-11-03 19:45:47,801] INFO [GroupCoordinator 0]: Assignment received from leader rdkafka-cf033a01-fd53-46c7-9ac9-d92cbfe8c9d8 for group test-group for generation 3. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:26:31,925] INFO Registered broker 0 at path /brokers/ids/0 with addresses: PLAINTEXT://localhost:9092, czxid (broker epoch): 25 (kafka.zk.KafkaZkClient) +[2023-11-06 13:26:31,966] INFO [ExpirationReaper-0-topic]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) +[2023-11-06 13:26:31,971] INFO [ExpirationReaper-0-Heartbeat]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) +[2023-11-06 13:26:31,972] INFO [ExpirationReaper-0-Rebalance]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) +[2023-11-06 13:26:31,972] INFO Successfully created /controller_epoch with initial epoch 0 (kafka.zk.KafkaZkClient) +[2023-11-06 13:26:31,981] INFO Feature ZK node created at path: /feature (kafka.server.FinalizedFeatureChangeListener) +[2023-11-06 13:26:31,982] INFO [GroupCoordinator 0]: Starting up. (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:26:31,986] INFO [GroupCoordinator 0]: Startup complete. (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:26:31,995] INFO [TransactionCoordinator id=0] Starting up. (kafka.coordinator.transaction.TransactionCoordinator) +[2023-11-06 13:26:31,997] INFO [MetadataCache brokerId=0] Updated cache from existing None to latest Features(version=3.6-IV2, finalizedFeatures={}, finalizedFeaturesEpoch=0). (kafka.server.metadata.ZkMetadataCache) +[2023-11-06 13:26:31,998] INFO [TxnMarkerSenderThread-0]: Starting (kafka.coordinator.transaction.TransactionMarkerChannelManager) +[2023-11-06 13:26:31,998] INFO [TransactionCoordinator id=0] Startup complete. (kafka.coordinator.transaction.TransactionCoordinator) +[2023-11-06 13:26:32,022] INFO [ExpirationReaper-0-AlterAcls]: Starting (kafka.server.DelayedOperationPurgatory$ExpiredOperationReaper) +[2023-11-06 13:26:32,049] INFO [/config/changes-event-process-thread]: Starting (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread) +[2023-11-06 13:26:32,051] INFO [Controller id=0, targetBrokerId=0] Node 0 disconnected. (org.apache.kafka.clients.NetworkClient) +[2023-11-06 13:26:32,053] WARN [Controller id=0, targetBrokerId=0] Connection to node 0 (localhost/127.0.0.1:9092) could not be established. Broker may not be available. (org.apache.kafka.clients.NetworkClient) +[2023-11-06 13:26:32,055] INFO [Controller id=0, targetBrokerId=0] Client requested connection close from node 0 (org.apache.kafka.clients.NetworkClient) +[2023-11-06 13:26:32,056] INFO [SocketServer listenerType=ZK_BROKER, nodeId=0] Enabling request processing. (kafka.network.SocketServer) +[2023-11-06 13:26:32,058] INFO Awaiting socket connections on 0.0.0.0:9092. (kafka.network.DataPlaneAcceptor) +[2023-11-06 13:26:32,063] INFO Kafka version: 3.6.0 (org.apache.kafka.common.utils.AppInfoParser) +[2023-11-06 13:26:32,063] INFO Kafka commitId: 60e845626d8a465a (org.apache.kafka.common.utils.AppInfoParser) +[2023-11-06 13:26:32,063] INFO Kafka startTimeMs: 1699295192060 (org.apache.kafka.common.utils.AppInfoParser) +[2023-11-06 13:26:32,064] INFO [KafkaServer id=0] started (kafka.server.KafkaServer) +[2023-11-06 13:26:32,273] INFO [zk-broker-0-to-controller-alter-partition-channel-manager]: Recorded new controller, from now on will use node localhost:9092 (id: 0 rack: null) (kafka.server.BrokerToControllerRequestThread) +[2023-11-06 13:26:32,278] INFO [zk-broker-0-to-controller-forwarding-channel-manager]: Recorded new controller, from now on will use node localhost:9092 (id: 0 rack: null) (kafka.server.BrokerToControllerRequestThread) +[2023-11-06 13:27:16,491] INFO Creating topic __consumer_offsets with configuration {compression.type=producer, cleanup.policy=compact, segment.bytes=104857600} and initial partition assignment HashMap(0 -> ArrayBuffer(0), 1 -> ArrayBuffer(0), 2 -> ArrayBuffer(0), 3 -> ArrayBuffer(0), 4 -> ArrayBuffer(0), 5 -> ArrayBuffer(0), 6 -> ArrayBuffer(0), 7 -> ArrayBuffer(0), 8 -> ArrayBuffer(0), 9 -> ArrayBuffer(0), 10 -> ArrayBuffer(0), 11 -> ArrayBuffer(0), 12 -> ArrayBuffer(0), 13 -> ArrayBuffer(0), 14 -> ArrayBuffer(0), 15 -> ArrayBuffer(0), 16 -> ArrayBuffer(0), 17 -> ArrayBuffer(0), 18 -> ArrayBuffer(0), 19 -> ArrayBuffer(0), 20 -> ArrayBuffer(0), 21 -> ArrayBuffer(0), 22 -> ArrayBuffer(0), 23 -> ArrayBuffer(0), 24 -> ArrayBuffer(0), 25 -> ArrayBuffer(0), 26 -> ArrayBuffer(0), 27 -> ArrayBuffer(0), 28 -> ArrayBuffer(0), 29 -> ArrayBuffer(0), 30 -> ArrayBuffer(0), 31 -> ArrayBuffer(0), 32 -> ArrayBuffer(0), 33 -> ArrayBuffer(0), 34 -> ArrayBuffer(0), 35 -> ArrayBuffer(0), 36 -> ArrayBuffer(0), 37 -> ArrayBuffer(0), 38 -> ArrayBuffer(0), 39 -> ArrayBuffer(0), 40 -> ArrayBuffer(0), 41 -> ArrayBuffer(0), 42 -> ArrayBuffer(0), 43 -> ArrayBuffer(0), 44 -> ArrayBuffer(0), 45 -> ArrayBuffer(0), 46 -> ArrayBuffer(0), 47 -> ArrayBuffer(0), 48 -> ArrayBuffer(0), 49 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient) +[2023-11-06 13:27:16,613] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions HashSet(__consumer_offsets-22, __consumer_offsets-30, __consumer_offsets-25, __consumer_offsets-35, __consumer_offsets-37, __consumer_offsets-38, __consumer_offsets-13, __consumer_offsets-8, __consumer_offsets-21, __consumer_offsets-4, __consumer_offsets-27, __consumer_offsets-7, __consumer_offsets-9, __consumer_offsets-46, __consumer_offsets-41, __consumer_offsets-33, __consumer_offsets-23, __consumer_offsets-49, __consumer_offsets-47, __consumer_offsets-16, __consumer_offsets-28, __consumer_offsets-31, __consumer_offsets-36, __consumer_offsets-42, __consumer_offsets-3, __consumer_offsets-18, __consumer_offsets-15, __consumer_offsets-24, __consumer_offsets-17, __consumer_offsets-48, __consumer_offsets-19, __consumer_offsets-11, __consumer_offsets-2, __consumer_offsets-43, __consumer_offsets-6, __consumer_offsets-14, __consumer_offsets-20, __consumer_offsets-0, __consumer_offsets-44, __consumer_offsets-39, __consumer_offsets-12, __consumer_offsets-45, __consumer_offsets-1, __consumer_offsets-5, __consumer_offsets-26, __consumer_offsets-29, __consumer_offsets-34, __consumer_offsets-10, __consumer_offsets-32, __consumer_offsets-40) (kafka.server.ReplicaFetcherManager) +[2023-11-06 13:27:16,652] INFO [LogLoader partition=__consumer_offsets-3, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,662] INFO Created log for partition __consumer_offsets-3 in /tmp/kafka-logs/__consumer_offsets-3 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,662] INFO [Partition __consumer_offsets-3 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-3 (kafka.cluster.Partition) +[2023-11-06 13:27:16,663] INFO [Partition __consumer_offsets-3 broker=0] Log loaded for partition __consumer_offsets-3 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,671] INFO [LogLoader partition=__consumer_offsets-18, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,672] INFO Created log for partition __consumer_offsets-18 in /tmp/kafka-logs/__consumer_offsets-18 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,672] INFO [Partition __consumer_offsets-18 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-18 (kafka.cluster.Partition) +[2023-11-06 13:27:16,672] INFO [Partition __consumer_offsets-18 broker=0] Log loaded for partition __consumer_offsets-18 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,675] INFO [LogLoader partition=__consumer_offsets-41, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,675] INFO Created log for partition __consumer_offsets-41 in /tmp/kafka-logs/__consumer_offsets-41 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,675] INFO [Partition __consumer_offsets-41 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-41 (kafka.cluster.Partition) +[2023-11-06 13:27:16,675] INFO [Partition __consumer_offsets-41 broker=0] Log loaded for partition __consumer_offsets-41 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,678] INFO [LogLoader partition=__consumer_offsets-10, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,679] INFO Created log for partition __consumer_offsets-10 in /tmp/kafka-logs/__consumer_offsets-10 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,679] INFO [Partition __consumer_offsets-10 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-10 (kafka.cluster.Partition) +[2023-11-06 13:27:16,679] INFO [Partition __consumer_offsets-10 broker=0] Log loaded for partition __consumer_offsets-10 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,681] INFO [LogLoader partition=__consumer_offsets-33, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,682] INFO Created log for partition __consumer_offsets-33 in /tmp/kafka-logs/__consumer_offsets-33 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,682] INFO [Partition __consumer_offsets-33 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-33 (kafka.cluster.Partition) +[2023-11-06 13:27:16,682] INFO [Partition __consumer_offsets-33 broker=0] Log loaded for partition __consumer_offsets-33 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,685] INFO [LogLoader partition=__consumer_offsets-48, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,685] INFO Created log for partition __consumer_offsets-48 in /tmp/kafka-logs/__consumer_offsets-48 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,686] INFO [Partition __consumer_offsets-48 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-48 (kafka.cluster.Partition) +[2023-11-06 13:27:16,686] INFO [Partition __consumer_offsets-48 broker=0] Log loaded for partition __consumer_offsets-48 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,688] INFO [LogLoader partition=__consumer_offsets-19, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,689] INFO Created log for partition __consumer_offsets-19 in /tmp/kafka-logs/__consumer_offsets-19 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,689] INFO [Partition __consumer_offsets-19 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-19 (kafka.cluster.Partition) +[2023-11-06 13:27:16,689] INFO [Partition __consumer_offsets-19 broker=0] Log loaded for partition __consumer_offsets-19 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,692] INFO [LogLoader partition=__consumer_offsets-34, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,692] INFO Created log for partition __consumer_offsets-34 in /tmp/kafka-logs/__consumer_offsets-34 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,692] INFO [Partition __consumer_offsets-34 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-34 (kafka.cluster.Partition) +[2023-11-06 13:27:16,693] INFO [Partition __consumer_offsets-34 broker=0] Log loaded for partition __consumer_offsets-34 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,695] INFO [LogLoader partition=__consumer_offsets-4, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,696] INFO Created log for partition __consumer_offsets-4 in /tmp/kafka-logs/__consumer_offsets-4 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,696] INFO [Partition __consumer_offsets-4 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-4 (kafka.cluster.Partition) +[2023-11-06 13:27:16,696] INFO [Partition __consumer_offsets-4 broker=0] Log loaded for partition __consumer_offsets-4 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,699] INFO [LogLoader partition=__consumer_offsets-11, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,699] INFO Created log for partition __consumer_offsets-11 in /tmp/kafka-logs/__consumer_offsets-11 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,699] INFO [Partition __consumer_offsets-11 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-11 (kafka.cluster.Partition) +[2023-11-06 13:27:16,699] INFO [Partition __consumer_offsets-11 broker=0] Log loaded for partition __consumer_offsets-11 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,702] INFO [LogLoader partition=__consumer_offsets-26, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,703] INFO Created log for partition __consumer_offsets-26 in /tmp/kafka-logs/__consumer_offsets-26 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,703] INFO [Partition __consumer_offsets-26 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-26 (kafka.cluster.Partition) +[2023-11-06 13:27:16,703] INFO [Partition __consumer_offsets-26 broker=0] Log loaded for partition __consumer_offsets-26 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,706] INFO [LogLoader partition=__consumer_offsets-49, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,706] INFO Created log for partition __consumer_offsets-49 in /tmp/kafka-logs/__consumer_offsets-49 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,706] INFO [Partition __consumer_offsets-49 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-49 (kafka.cluster.Partition) +[2023-11-06 13:27:16,706] INFO [Partition __consumer_offsets-49 broker=0] Log loaded for partition __consumer_offsets-49 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,709] INFO [LogLoader partition=__consumer_offsets-39, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,710] INFO Created log for partition __consumer_offsets-39 in /tmp/kafka-logs/__consumer_offsets-39 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,710] INFO [Partition __consumer_offsets-39 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-39 (kafka.cluster.Partition) +[2023-11-06 13:27:16,710] INFO [Partition __consumer_offsets-39 broker=0] Log loaded for partition __consumer_offsets-39 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,713] INFO [LogLoader partition=__consumer_offsets-9, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,713] INFO Created log for partition __consumer_offsets-9 in /tmp/kafka-logs/__consumer_offsets-9 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,713] INFO [Partition __consumer_offsets-9 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-9 (kafka.cluster.Partition) +[2023-11-06 13:27:16,713] INFO [Partition __consumer_offsets-9 broker=0] Log loaded for partition __consumer_offsets-9 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,716] INFO [LogLoader partition=__consumer_offsets-24, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,716] INFO Created log for partition __consumer_offsets-24 in /tmp/kafka-logs/__consumer_offsets-24 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,716] INFO [Partition __consumer_offsets-24 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-24 (kafka.cluster.Partition) +[2023-11-06 13:27:16,716] INFO [Partition __consumer_offsets-24 broker=0] Log loaded for partition __consumer_offsets-24 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,720] INFO [LogLoader partition=__consumer_offsets-31, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,720] INFO Created log for partition __consumer_offsets-31 in /tmp/kafka-logs/__consumer_offsets-31 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,720] INFO [Partition __consumer_offsets-31 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-31 (kafka.cluster.Partition) +[2023-11-06 13:27:16,720] INFO [Partition __consumer_offsets-31 broker=0] Log loaded for partition __consumer_offsets-31 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,723] INFO [LogLoader partition=__consumer_offsets-46, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,723] INFO Created log for partition __consumer_offsets-46 in /tmp/kafka-logs/__consumer_offsets-46 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,724] INFO [Partition __consumer_offsets-46 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-46 (kafka.cluster.Partition) +[2023-11-06 13:27:16,724] INFO [Partition __consumer_offsets-46 broker=0] Log loaded for partition __consumer_offsets-46 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,727] INFO [LogLoader partition=__consumer_offsets-1, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,727] INFO Created log for partition __consumer_offsets-1 in /tmp/kafka-logs/__consumer_offsets-1 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,727] INFO [Partition __consumer_offsets-1 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-1 (kafka.cluster.Partition) +[2023-11-06 13:27:16,727] INFO [Partition __consumer_offsets-1 broker=0] Log loaded for partition __consumer_offsets-1 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,731] INFO [LogLoader partition=__consumer_offsets-16, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,731] INFO Created log for partition __consumer_offsets-16 in /tmp/kafka-logs/__consumer_offsets-16 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,731] INFO [Partition __consumer_offsets-16 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-16 (kafka.cluster.Partition) +[2023-11-06 13:27:16,731] INFO [Partition __consumer_offsets-16 broker=0] Log loaded for partition __consumer_offsets-16 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,736] INFO [LogLoader partition=__consumer_offsets-2, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,736] INFO Created log for partition __consumer_offsets-2 in /tmp/kafka-logs/__consumer_offsets-2 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,736] INFO [Partition __consumer_offsets-2 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-2 (kafka.cluster.Partition) +[2023-11-06 13:27:16,736] INFO [Partition __consumer_offsets-2 broker=0] Log loaded for partition __consumer_offsets-2 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,740] INFO [LogLoader partition=__consumer_offsets-25, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,740] INFO Created log for partition __consumer_offsets-25 in /tmp/kafka-logs/__consumer_offsets-25 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,740] INFO [Partition __consumer_offsets-25 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-25 (kafka.cluster.Partition) +[2023-11-06 13:27:16,740] INFO [Partition __consumer_offsets-25 broker=0] Log loaded for partition __consumer_offsets-25 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,744] INFO [LogLoader partition=__consumer_offsets-40, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,744] INFO Created log for partition __consumer_offsets-40 in /tmp/kafka-logs/__consumer_offsets-40 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,744] INFO [Partition __consumer_offsets-40 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-40 (kafka.cluster.Partition) +[2023-11-06 13:27:16,744] INFO [Partition __consumer_offsets-40 broker=0] Log loaded for partition __consumer_offsets-40 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,748] INFO [LogLoader partition=__consumer_offsets-47, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,748] INFO Created log for partition __consumer_offsets-47 in /tmp/kafka-logs/__consumer_offsets-47 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,748] INFO [Partition __consumer_offsets-47 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-47 (kafka.cluster.Partition) +[2023-11-06 13:27:16,748] INFO [Partition __consumer_offsets-47 broker=0] Log loaded for partition __consumer_offsets-47 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,752] INFO [LogLoader partition=__consumer_offsets-17, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,753] INFO Created log for partition __consumer_offsets-17 in /tmp/kafka-logs/__consumer_offsets-17 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,753] INFO [Partition __consumer_offsets-17 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-17 (kafka.cluster.Partition) +[2023-11-06 13:27:16,753] INFO [Partition __consumer_offsets-17 broker=0] Log loaded for partition __consumer_offsets-17 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,756] INFO [LogLoader partition=__consumer_offsets-32, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,757] INFO Created log for partition __consumer_offsets-32 in /tmp/kafka-logs/__consumer_offsets-32 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,757] INFO [Partition __consumer_offsets-32 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-32 (kafka.cluster.Partition) +[2023-11-06 13:27:16,757] INFO [Partition __consumer_offsets-32 broker=0] Log loaded for partition __consumer_offsets-32 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,759] INFO [LogLoader partition=__consumer_offsets-37, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,760] INFO Created log for partition __consumer_offsets-37 in /tmp/kafka-logs/__consumer_offsets-37 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,760] INFO [Partition __consumer_offsets-37 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-37 (kafka.cluster.Partition) +[2023-11-06 13:27:16,760] INFO [Partition __consumer_offsets-37 broker=0] Log loaded for partition __consumer_offsets-37 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,763] INFO [LogLoader partition=__consumer_offsets-7, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,763] INFO Created log for partition __consumer_offsets-7 in /tmp/kafka-logs/__consumer_offsets-7 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,763] INFO [Partition __consumer_offsets-7 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-7 (kafka.cluster.Partition) +[2023-11-06 13:27:16,763] INFO [Partition __consumer_offsets-7 broker=0] Log loaded for partition __consumer_offsets-7 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,766] INFO [LogLoader partition=__consumer_offsets-22, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,766] INFO Created log for partition __consumer_offsets-22 in /tmp/kafka-logs/__consumer_offsets-22 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,766] INFO [Partition __consumer_offsets-22 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-22 (kafka.cluster.Partition) +[2023-11-06 13:27:16,766] INFO [Partition __consumer_offsets-22 broker=0] Log loaded for partition __consumer_offsets-22 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,769] INFO [LogLoader partition=__consumer_offsets-29, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,770] INFO Created log for partition __consumer_offsets-29 in /tmp/kafka-logs/__consumer_offsets-29 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,770] INFO [Partition __consumer_offsets-29 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-29 (kafka.cluster.Partition) +[2023-11-06 13:27:16,770] INFO [Partition __consumer_offsets-29 broker=0] Log loaded for partition __consumer_offsets-29 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,773] INFO [LogLoader partition=__consumer_offsets-44, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,774] INFO Created log for partition __consumer_offsets-44 in /tmp/kafka-logs/__consumer_offsets-44 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,774] INFO [Partition __consumer_offsets-44 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-44 (kafka.cluster.Partition) +[2023-11-06 13:27:16,774] INFO [Partition __consumer_offsets-44 broker=0] Log loaded for partition __consumer_offsets-44 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,777] INFO [LogLoader partition=__consumer_offsets-14, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,778] INFO Created log for partition __consumer_offsets-14 in /tmp/kafka-logs/__consumer_offsets-14 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,778] INFO [Partition __consumer_offsets-14 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-14 (kafka.cluster.Partition) +[2023-11-06 13:27:16,778] INFO [Partition __consumer_offsets-14 broker=0] Log loaded for partition __consumer_offsets-14 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,781] INFO [LogLoader partition=__consumer_offsets-23, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,781] INFO Created log for partition __consumer_offsets-23 in /tmp/kafka-logs/__consumer_offsets-23 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,781] INFO [Partition __consumer_offsets-23 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-23 (kafka.cluster.Partition) +[2023-11-06 13:27:16,781] INFO [Partition __consumer_offsets-23 broker=0] Log loaded for partition __consumer_offsets-23 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,784] INFO [LogLoader partition=__consumer_offsets-38, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,784] INFO Created log for partition __consumer_offsets-38 in /tmp/kafka-logs/__consumer_offsets-38 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,784] INFO [Partition __consumer_offsets-38 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-38 (kafka.cluster.Partition) +[2023-11-06 13:27:16,784] INFO [Partition __consumer_offsets-38 broker=0] Log loaded for partition __consumer_offsets-38 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,787] INFO [LogLoader partition=__consumer_offsets-8, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,787] INFO Created log for partition __consumer_offsets-8 in /tmp/kafka-logs/__consumer_offsets-8 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,787] INFO [Partition __consumer_offsets-8 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-8 (kafka.cluster.Partition) +[2023-11-06 13:27:16,788] INFO [Partition __consumer_offsets-8 broker=0] Log loaded for partition __consumer_offsets-8 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,791] INFO [LogLoader partition=__consumer_offsets-45, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,791] INFO Created log for partition __consumer_offsets-45 in /tmp/kafka-logs/__consumer_offsets-45 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,792] INFO [Partition __consumer_offsets-45 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-45 (kafka.cluster.Partition) +[2023-11-06 13:27:16,792] INFO [Partition __consumer_offsets-45 broker=0] Log loaded for partition __consumer_offsets-45 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,795] INFO [LogLoader partition=__consumer_offsets-15, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,795] INFO Created log for partition __consumer_offsets-15 in /tmp/kafka-logs/__consumer_offsets-15 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,795] INFO [Partition __consumer_offsets-15 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-15 (kafka.cluster.Partition) +[2023-11-06 13:27:16,796] INFO [Partition __consumer_offsets-15 broker=0] Log loaded for partition __consumer_offsets-15 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,798] INFO [LogLoader partition=__consumer_offsets-30, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,799] INFO Created log for partition __consumer_offsets-30 in /tmp/kafka-logs/__consumer_offsets-30 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,799] INFO [Partition __consumer_offsets-30 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-30 (kafka.cluster.Partition) +[2023-11-06 13:27:16,799] INFO [Partition __consumer_offsets-30 broker=0] Log loaded for partition __consumer_offsets-30 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,802] INFO [LogLoader partition=__consumer_offsets-0, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,803] INFO Created log for partition __consumer_offsets-0 in /tmp/kafka-logs/__consumer_offsets-0 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,803] INFO [Partition __consumer_offsets-0 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,803] INFO [Partition __consumer_offsets-0 broker=0] Log loaded for partition __consumer_offsets-0 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,806] INFO [LogLoader partition=__consumer_offsets-35, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,806] INFO Created log for partition __consumer_offsets-35 in /tmp/kafka-logs/__consumer_offsets-35 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,806] INFO [Partition __consumer_offsets-35 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-35 (kafka.cluster.Partition) +[2023-11-06 13:27:16,807] INFO [Partition __consumer_offsets-35 broker=0] Log loaded for partition __consumer_offsets-35 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,810] INFO [LogLoader partition=__consumer_offsets-5, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,810] INFO Created log for partition __consumer_offsets-5 in /tmp/kafka-logs/__consumer_offsets-5 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,810] INFO [Partition __consumer_offsets-5 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-5 (kafka.cluster.Partition) +[2023-11-06 13:27:16,810] INFO [Partition __consumer_offsets-5 broker=0] Log loaded for partition __consumer_offsets-5 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,813] INFO [LogLoader partition=__consumer_offsets-20, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,814] INFO Created log for partition __consumer_offsets-20 in /tmp/kafka-logs/__consumer_offsets-20 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,814] INFO [Partition __consumer_offsets-20 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-20 (kafka.cluster.Partition) +[2023-11-06 13:27:16,814] INFO [Partition __consumer_offsets-20 broker=0] Log loaded for partition __consumer_offsets-20 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,817] INFO [LogLoader partition=__consumer_offsets-27, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,817] INFO Created log for partition __consumer_offsets-27 in /tmp/kafka-logs/__consumer_offsets-27 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,817] INFO [Partition __consumer_offsets-27 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-27 (kafka.cluster.Partition) +[2023-11-06 13:27:16,817] INFO [Partition __consumer_offsets-27 broker=0] Log loaded for partition __consumer_offsets-27 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,821] INFO [LogLoader partition=__consumer_offsets-42, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,821] INFO Created log for partition __consumer_offsets-42 in /tmp/kafka-logs/__consumer_offsets-42 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,821] INFO [Partition __consumer_offsets-42 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-42 (kafka.cluster.Partition) +[2023-11-06 13:27:16,821] INFO [Partition __consumer_offsets-42 broker=0] Log loaded for partition __consumer_offsets-42 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,825] INFO [LogLoader partition=__consumer_offsets-12, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,825] INFO Created log for partition __consumer_offsets-12 in /tmp/kafka-logs/__consumer_offsets-12 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,825] INFO [Partition __consumer_offsets-12 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-12 (kafka.cluster.Partition) +[2023-11-06 13:27:16,825] INFO [Partition __consumer_offsets-12 broker=0] Log loaded for partition __consumer_offsets-12 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,828] INFO [LogLoader partition=__consumer_offsets-21, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,829] INFO Created log for partition __consumer_offsets-21 in /tmp/kafka-logs/__consumer_offsets-21 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,829] INFO [Partition __consumer_offsets-21 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-21 (kafka.cluster.Partition) +[2023-11-06 13:27:16,829] INFO [Partition __consumer_offsets-21 broker=0] Log loaded for partition __consumer_offsets-21 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,832] INFO [LogLoader partition=__consumer_offsets-36, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,832] INFO Created log for partition __consumer_offsets-36 in /tmp/kafka-logs/__consumer_offsets-36 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,832] INFO [Partition __consumer_offsets-36 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-36 (kafka.cluster.Partition) +[2023-11-06 13:27:16,832] INFO [Partition __consumer_offsets-36 broker=0] Log loaded for partition __consumer_offsets-36 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,835] INFO [LogLoader partition=__consumer_offsets-6, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,836] INFO Created log for partition __consumer_offsets-6 in /tmp/kafka-logs/__consumer_offsets-6 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,836] INFO [Partition __consumer_offsets-6 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-6 (kafka.cluster.Partition) +[2023-11-06 13:27:16,836] INFO [Partition __consumer_offsets-6 broker=0] Log loaded for partition __consumer_offsets-6 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,839] INFO [LogLoader partition=__consumer_offsets-43, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,840] INFO Created log for partition __consumer_offsets-43 in /tmp/kafka-logs/__consumer_offsets-43 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,840] INFO [Partition __consumer_offsets-43 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-43 (kafka.cluster.Partition) +[2023-11-06 13:27:16,840] INFO [Partition __consumer_offsets-43 broker=0] Log loaded for partition __consumer_offsets-43 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,843] INFO [LogLoader partition=__consumer_offsets-13, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,844] INFO Created log for partition __consumer_offsets-13 in /tmp/kafka-logs/__consumer_offsets-13 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,844] INFO [Partition __consumer_offsets-13 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-13 (kafka.cluster.Partition) +[2023-11-06 13:27:16,844] INFO [Partition __consumer_offsets-13 broker=0] Log loaded for partition __consumer_offsets-13 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,846] INFO [LogLoader partition=__consumer_offsets-28, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:16,846] INFO Created log for partition __consumer_offsets-28 in /tmp/kafka-logs/__consumer_offsets-28 with properties {cleanup.policy=compact, compression.type="producer", segment.bytes=104857600} (kafka.log.LogManager) +[2023-11-06 13:27:16,846] INFO [Partition __consumer_offsets-28 broker=0] No checkpointed highwatermark is found for partition __consumer_offsets-28 (kafka.cluster.Partition) +[2023-11-06 13:27:16,847] INFO [Partition __consumer_offsets-28 broker=0] Log loaded for partition __consumer_offsets-28 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:16,851] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 3 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,851] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-3 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 18 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-18 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 41 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-41 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 10 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-10 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 33 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-33 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 48 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-48 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 19 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-19 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 34 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-34 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 4 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-4 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 11 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-11 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 26 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-26 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 49 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-49 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 39 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-39 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 9 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-9 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 24 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-24 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 31 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-31 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 46 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-46 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 1 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-1 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 16 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-16 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 2 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-2 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 25 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-25 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 40 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-40 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 47 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-47 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 17 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-17 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 32 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-32 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 37 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-37 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 7 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-7 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 22 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-22 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 29 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-29 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 44 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-44 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 14 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-14 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 23 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,852] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-23 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,852] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 38 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-38 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 8 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-8 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 45 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-45 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 15 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-15 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 30 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-30 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 0 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-0 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 35 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-35 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 5 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-5 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 20 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-20 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 27 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-27 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 42 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-42 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 12 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-12 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 21 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-21 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 36 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-36 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 6 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-6 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 43 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-43 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 13 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-13 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,853] INFO [GroupCoordinator 0]: Elected as the group coordinator for partition 28 in epoch 0 (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:16,853] INFO [GroupMetadataManager brokerId=0] Scheduling loading of offsets and group metadata from __consumer_offsets-28 for epoch 0 (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-3 in 3 milliseconds for epoch 0, of which 1 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-18 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-41 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-10 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-33 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-48 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-19 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-34 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-4 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-11 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,855] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-26 in 3 milliseconds for epoch 0, of which 3 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-49 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-39 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-9 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-24 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-31 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-46 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-1 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-16 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-2 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-25 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-40 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-47 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-17 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,856] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-32 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-37 in 5 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-7 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-22 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-29 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-44 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-14 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-23 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-38 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-8 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-45 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-15 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-30 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,857] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-0 in 4 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-35 in 5 milliseconds for epoch 0, of which 4 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-5 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-20 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-27 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-42 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-12 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-21 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-36 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-6 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-43 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,858] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-13 in 5 milliseconds for epoch 0, of which 5 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:16,859] INFO [GroupMetadataManager brokerId=0] Finished loading offsets and group metadata from __consumer_offsets-28 in 6 milliseconds for epoch 0, of which 6 milliseconds was spent in the scheduler. (kafka.coordinator.group.GroupMetadataManager) +[2023-11-06 13:27:32,123] INFO Creating topic test-topic with configuration {} and initial partition assignment HashMap(0 -> ArrayBuffer(0)) (kafka.zk.AdminZkClient) +[2023-11-06 13:27:32,133] INFO [ReplicaFetcherManager on broker 0] Removed fetcher for partitions Set(test-topic-0) (kafka.server.ReplicaFetcherManager) +[2023-11-06 13:27:32,134] INFO [LogLoader partition=test-topic-0, dir=/tmp/kafka-logs] Loading producer state till offset 0 with message format version 2 (kafka.log.UnifiedLog$) +[2023-11-06 13:27:32,135] INFO Created log for partition test-topic-0 in /tmp/kafka-logs/test-topic-0 with properties {} (kafka.log.LogManager) +[2023-11-06 13:27:32,136] INFO [Partition test-topic-0 broker=0] No checkpointed highwatermark is found for partition test-topic-0 (kafka.cluster.Partition) +[2023-11-06 13:27:32,136] INFO [Partition test-topic-0 broker=0] Log loaded for partition test-topic-0 with initial high watermark 0 (kafka.cluster.Partition) +[2023-11-06 13:27:34,588] INFO [GroupCoordinator 0]: Dynamic member with unknown member id joins group test-group in Empty state. Created a new member id rdkafka-ca297002-849e-4f49-bd14-13700d89fb0f and request the member to rejoin with this id. (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:34,594] INFO [GroupCoordinator 0]: Preparing to rebalance group test-group in state PreparingRebalance with old generation 0 (__consumer_offsets-12) (reason: Adding new member rdkafka-ca297002-849e-4f49-bd14-13700d89fb0f with group instance id None; client reason: not provided) (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:34,608] INFO [GroupCoordinator 0]: Stabilized group test-group generation 1 (__consumer_offsets-12) with 1 members (kafka.coordinator.group.GroupCoordinator) +[2023-11-06 13:27:34,614] INFO [GroupCoordinator 0]: Assignment received from leader rdkafka-ca297002-849e-4f49-bd14-13700d89fb0f for group test-group for generation 1. The group has 1 members, 0 of which are static. (kafka.coordinator.group.GroupCoordinator) diff --git a/logs/state-change.log b/logs/state-change.log index 14e0dc6..c5dfb2f 100644 --- a/logs/state-change.log +++ b/logs/state-change.log @@ -1,231 +1,172 @@ -[2023-11-03 19:38:21,217] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 0 partitions (state.change.logger) -[2023-11-03 19:38:41,912] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-22 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,912] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-30 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,912] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-25 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,912] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-35 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-37 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-38 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-13 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-8 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-21 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-4 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-27 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-7 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-9 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-46 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-41 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-33 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-23 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-49 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-47 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-16 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-28 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-31 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-36 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-42 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-3 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-18 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-15 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-24 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-17 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-48 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-19 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-11 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-2 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-43 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-6 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-14 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-20 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,913] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-44 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-39 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-12 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-45 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-1 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-5 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-26 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-29 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-34 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-10 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-32 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-40 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:41,914] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) -[2023-11-03 19:38:41,917] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-22 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-30 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-25 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-35 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-37 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-38 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-13 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-8 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-21 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-4 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-27 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-7 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-9 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-46 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-41 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-33 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-23 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-49 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-47 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-16 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-28 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-31 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-36 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-42 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-3 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,973] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-18 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-15 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-24 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-17 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-48 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-19 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-11 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-2 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-43 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-6 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-14 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-20 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-44 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,974] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-39 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,975] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-12 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,975] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-45 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,975] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-1 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,975] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-5 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,975] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-26 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,975] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-29 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,975] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-34 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,975] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-10 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,975] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-32 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,975] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-40 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:41,978] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 50 become-leader and 0 become-follower partitions (state.change.logger) -[2023-11-03 19:38:41,980] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 50 partitions (state.change.logger) -[2023-11-03 19:38:41,981] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) -[2023-11-03 19:38:41,985] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 1 from controller 0 for 50 partitions (state.change.logger) -[2023-11-03 19:38:42,030] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 1 from controller 0 epoch 1 as part of the become-leader transition for 50 partitions (state.change.logger) -[2023-11-03 19:38:42,082] INFO [Broker id=0] Leader __consumer_offsets-3 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,089] INFO [Broker id=0] Leader __consumer_offsets-18 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,096] INFO [Broker id=0] Leader __consumer_offsets-41 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,100] INFO [Broker id=0] Leader __consumer_offsets-10 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,104] INFO [Broker id=0] Leader __consumer_offsets-33 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,108] INFO [Broker id=0] Leader __consumer_offsets-48 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,114] INFO [Broker id=0] Leader __consumer_offsets-19 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,118] INFO [Broker id=0] Leader __consumer_offsets-34 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,122] INFO [Broker id=0] Leader __consumer_offsets-4 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,127] INFO [Broker id=0] Leader __consumer_offsets-11 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,131] INFO [Broker id=0] Leader __consumer_offsets-26 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,135] INFO [Broker id=0] Leader __consumer_offsets-49 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,138] INFO [Broker id=0] Leader __consumer_offsets-39 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,143] INFO [Broker id=0] Leader __consumer_offsets-9 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,146] INFO [Broker id=0] Leader __consumer_offsets-24 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,149] INFO [Broker id=0] Leader __consumer_offsets-31 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,153] INFO [Broker id=0] Leader __consumer_offsets-46 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,156] INFO [Broker id=0] Leader __consumer_offsets-1 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,163] INFO [Broker id=0] Leader __consumer_offsets-16 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,167] INFO [Broker id=0] Leader __consumer_offsets-2 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,171] INFO [Broker id=0] Leader __consumer_offsets-25 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,176] INFO [Broker id=0] Leader __consumer_offsets-40 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,181] INFO [Broker id=0] Leader __consumer_offsets-47 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,185] INFO [Broker id=0] Leader __consumer_offsets-17 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,188] INFO [Broker id=0] Leader __consumer_offsets-32 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,195] INFO [Broker id=0] Leader __consumer_offsets-37 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,201] INFO [Broker id=0] Leader __consumer_offsets-7 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,204] INFO [Broker id=0] Leader __consumer_offsets-22 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,210] INFO [Broker id=0] Leader __consumer_offsets-29 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,216] INFO [Broker id=0] Leader __consumer_offsets-44 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,220] INFO [Broker id=0] Leader __consumer_offsets-14 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,226] INFO [Broker id=0] Leader __consumer_offsets-23 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,231] INFO [Broker id=0] Leader __consumer_offsets-38 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,236] INFO [Broker id=0] Leader __consumer_offsets-8 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,239] INFO [Broker id=0] Leader __consumer_offsets-45 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,244] INFO [Broker id=0] Leader __consumer_offsets-15 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,249] INFO [Broker id=0] Leader __consumer_offsets-30 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,252] INFO [Broker id=0] Leader __consumer_offsets-0 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,255] INFO [Broker id=0] Leader __consumer_offsets-35 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,261] INFO [Broker id=0] Leader __consumer_offsets-5 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,266] INFO [Broker id=0] Leader __consumer_offsets-20 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,270] INFO [Broker id=0] Leader __consumer_offsets-27 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,273] INFO [Broker id=0] Leader __consumer_offsets-42 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,277] INFO [Broker id=0] Leader __consumer_offsets-12 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,280] INFO [Broker id=0] Leader __consumer_offsets-21 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,285] INFO [Broker id=0] Leader __consumer_offsets-36 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,288] INFO [Broker id=0] Leader __consumer_offsets-6 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,292] INFO [Broker id=0] Leader __consumer_offsets-43 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,296] INFO [Broker id=0] Leader __consumer_offsets-13 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,299] INFO [Broker id=0] Leader __consumer_offsets-28 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:42,310] INFO [Broker id=0] Finished LeaderAndIsr request in 325ms correlationId 1 from controller 0 for 50 partitions (state.change.logger) -[2023-11-03 19:38:42,318] INFO [Broker id=0] Add 50 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 2 (state.change.logger) -[2023-11-03 19:38:56,520] INFO [Controller id=0 epoch=1] Changed partition test-topic-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) -[2023-11-03 19:38:56,520] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) -[2023-11-03 19:38:56,520] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) -[2023-11-03 19:38:56,526] INFO [Controller id=0 epoch=1] Changed partition test-topic-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) -[2023-11-03 19:38:56,526] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger) -[2023-11-03 19:38:56,526] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger) -[2023-11-03 19:38:56,526] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) -[2023-11-03 19:38:56,526] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 3 from controller 0 for 1 partitions (state.change.logger) -[2023-11-03 19:38:56,527] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 3 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger) -[2023-11-03 19:38:56,530] INFO [Broker id=0] Leader test-topic-0 with topic id Some(Pb9zfnlKRkmTGaMQyAABkw) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:38:56,530] INFO [Broker id=0] Finished LeaderAndIsr request in 4ms correlationId 3 from controller 0 for 1 partitions (state.change.logger) -[2023-11-03 19:38:56,532] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 4 (state.change.logger) -[2023-11-03 19:40:48,863] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) -[2023-11-03 19:45:29,526] INFO [Controller id=0 epoch=2] Sending UpdateMetadata request to brokers HashSet(0) for 0 partitions (state.change.logger) -[2023-11-03 19:45:29,562] INFO [Controller id=0 epoch=2] Sending LeaderAndIsr request to broker 0 with 51 become-leader and 0 become-follower partitions (state.change.logger) -[2023-11-03 19:45:29,564] INFO [Controller id=0 epoch=2] Sending UpdateMetadata request to brokers HashSet(0) for 51 partitions (state.change.logger) -[2023-11-03 19:45:29,686] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 1 from controller 0 for 51 partitions (state.change.logger) -[2023-11-03 19:45:29,711] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 1 from controller 0 epoch 2 as part of the become-leader transition for 51 partitions (state.change.logger) -[2023-11-03 19:45:29,719] INFO [Broker id=0] Leader __consumer_offsets-3 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,722] INFO [Broker id=0] Leader __consumer_offsets-18 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,722] INFO [Broker id=0] Leader __consumer_offsets-41 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,722] INFO [Broker id=0] Leader __consumer_offsets-10 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,722] INFO [Broker id=0] Leader __consumer_offsets-33 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,723] INFO [Broker id=0] Leader __consumer_offsets-48 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,723] INFO [Broker id=0] Leader __consumer_offsets-19 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,723] INFO [Broker id=0] Leader __consumer_offsets-34 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,724] INFO [Broker id=0] Leader __consumer_offsets-4 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,724] INFO [Broker id=0] Leader __consumer_offsets-11 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,724] INFO [Broker id=0] Leader __consumer_offsets-26 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,724] INFO [Broker id=0] Leader __consumer_offsets-49 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,725] INFO [Broker id=0] Leader __consumer_offsets-39 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,725] INFO [Broker id=0] Leader __consumer_offsets-9 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,726] INFO [Broker id=0] Leader __consumer_offsets-24 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,726] INFO [Broker id=0] Leader __consumer_offsets-31 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,726] INFO [Broker id=0] Leader __consumer_offsets-46 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,726] INFO [Broker id=0] Leader __consumer_offsets-1 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,727] INFO [Broker id=0] Leader __consumer_offsets-16 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,727] INFO [Broker id=0] Leader __consumer_offsets-2 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,728] INFO [Broker id=0] Leader __consumer_offsets-25 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,728] INFO [Broker id=0] Leader __consumer_offsets-40 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,729] INFO [Broker id=0] Leader __consumer_offsets-47 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,729] INFO [Broker id=0] Leader __consumer_offsets-17 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,729] INFO [Broker id=0] Leader __consumer_offsets-32 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,730] INFO [Broker id=0] Leader __consumer_offsets-37 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,730] INFO [Broker id=0] Leader __consumer_offsets-7 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,730] INFO [Broker id=0] Leader __consumer_offsets-22 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,730] INFO [Broker id=0] Leader __consumer_offsets-29 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,731] INFO [Broker id=0] Leader __consumer_offsets-44 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,731] INFO [Broker id=0] Leader __consumer_offsets-14 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,732] INFO [Broker id=0] Leader __consumer_offsets-23 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,732] INFO [Broker id=0] Leader __consumer_offsets-38 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,732] INFO [Broker id=0] Leader __consumer_offsets-8 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,732] INFO [Broker id=0] Leader __consumer_offsets-45 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,732] INFO [Broker id=0] Leader __consumer_offsets-15 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,733] INFO [Broker id=0] Leader __consumer_offsets-30 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,733] INFO [Broker id=0] Leader __consumer_offsets-0 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,733] INFO [Broker id=0] Leader __consumer_offsets-35 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,734] INFO [Broker id=0] Leader __consumer_offsets-5 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,734] INFO [Broker id=0] Leader __consumer_offsets-20 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,734] INFO [Broker id=0] Leader __consumer_offsets-27 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,734] INFO [Broker id=0] Leader __consumer_offsets-42 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,735] INFO [Broker id=0] Leader __consumer_offsets-12 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 7 with partition epoch 0, high watermark 7, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,735] INFO [Broker id=0] Leader __consumer_offsets-21 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,735] INFO [Broker id=0] Leader test-topic-0 with topic id Some(Pb9zfnlKRkmTGaMQyAABkw) starts at leader epoch 0 from offset 5 with partition epoch 0, high watermark 5, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,735] INFO [Broker id=0] Leader __consumer_offsets-36 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,736] INFO [Broker id=0] Leader __consumer_offsets-6 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,736] INFO [Broker id=0] Leader __consumer_offsets-43 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,736] INFO [Broker id=0] Leader __consumer_offsets-13 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,736] INFO [Broker id=0] Leader __consumer_offsets-28 with topic id Some(vfIIZeo7TYKNeTmKGzMlyg) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) -[2023-11-03 19:45:29,749] INFO [Broker id=0] Finished LeaderAndIsr request in 63ms correlationId 1 from controller 0 for 51 partitions (state.change.logger) -[2023-11-03 19:45:29,757] INFO [Broker id=0] Add 51 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 2 with correlation id 2 (state.change.logger) +[2023-11-06 13:26:32,032] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 0 partitions (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-22 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-30 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-25 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-35 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-37 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-38 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-13 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-8 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-21 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-4 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-27 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-7 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-9 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-46 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,513] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-41 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-33 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-23 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-49 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-47 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-16 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-28 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-31 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-36 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-42 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-3 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-18 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-15 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-24 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-17 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-48 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-19 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-11 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-2 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-43 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-6 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-14 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-20 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-44 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-39 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-12 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-45 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-1 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-5 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-26 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-29 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-34 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-10 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-32 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-40 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:16,514] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) +[2023-11-06 13:27:16,517] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) +[2023-11-06 13:27:16,564] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-22 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,564] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-30 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,564] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-25 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,564] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-35 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,564] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-37 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,564] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-38 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,564] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-13 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-8 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-21 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-4 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-27 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-7 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-9 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-46 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-41 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-33 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-23 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-49 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-47 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-16 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-28 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-31 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-36 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-42 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-3 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-18 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-15 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-24 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-17 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-48 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-19 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-11 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-2 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-43 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-6 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-14 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-20 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-44 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-39 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-12 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-45 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-1 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-5 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-26 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-29 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,565] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-34 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,566] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-10 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,566] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-32 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,566] INFO [Controller id=0 epoch=1] Changed partition __consumer_offsets-40 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:16,567] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 50 become-leader and 0 become-follower partitions (state.change.logger) +[2023-11-06 13:27:16,568] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 50 partitions (state.change.logger) +[2023-11-06 13:27:16,569] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) +[2023-11-06 13:27:16,574] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 1 from controller 0 for 50 partitions (state.change.logger) +[2023-11-06 13:27:16,614] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 1 from controller 0 epoch 1 as part of the become-leader transition for 50 partitions (state.change.logger) +[2023-11-06 13:27:16,665] INFO [Broker id=0] Leader __consumer_offsets-3 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,672] INFO [Broker id=0] Leader __consumer_offsets-18 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,675] INFO [Broker id=0] Leader __consumer_offsets-41 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,679] INFO [Broker id=0] Leader __consumer_offsets-10 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,682] INFO [Broker id=0] Leader __consumer_offsets-33 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,686] INFO [Broker id=0] Leader __consumer_offsets-48 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,689] INFO [Broker id=0] Leader __consumer_offsets-19 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,693] INFO [Broker id=0] Leader __consumer_offsets-34 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,696] INFO [Broker id=0] Leader __consumer_offsets-4 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,699] INFO [Broker id=0] Leader __consumer_offsets-11 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,703] INFO [Broker id=0] Leader __consumer_offsets-26 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,706] INFO [Broker id=0] Leader __consumer_offsets-49 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,710] INFO [Broker id=0] Leader __consumer_offsets-39 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,713] INFO [Broker id=0] Leader __consumer_offsets-9 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,716] INFO [Broker id=0] Leader __consumer_offsets-24 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,720] INFO [Broker id=0] Leader __consumer_offsets-31 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,724] INFO [Broker id=0] Leader __consumer_offsets-46 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,727] INFO [Broker id=0] Leader __consumer_offsets-1 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,732] INFO [Broker id=0] Leader __consumer_offsets-16 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,736] INFO [Broker id=0] Leader __consumer_offsets-2 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,740] INFO [Broker id=0] Leader __consumer_offsets-25 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,745] INFO [Broker id=0] Leader __consumer_offsets-40 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,749] INFO [Broker id=0] Leader __consumer_offsets-47 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,753] INFO [Broker id=0] Leader __consumer_offsets-17 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,757] INFO [Broker id=0] Leader __consumer_offsets-32 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,760] INFO [Broker id=0] Leader __consumer_offsets-37 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,764] INFO [Broker id=0] Leader __consumer_offsets-7 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,766] INFO [Broker id=0] Leader __consumer_offsets-22 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,770] INFO [Broker id=0] Leader __consumer_offsets-29 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,774] INFO [Broker id=0] Leader __consumer_offsets-44 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,778] INFO [Broker id=0] Leader __consumer_offsets-14 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,781] INFO [Broker id=0] Leader __consumer_offsets-23 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,784] INFO [Broker id=0] Leader __consumer_offsets-38 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,788] INFO [Broker id=0] Leader __consumer_offsets-8 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,792] INFO [Broker id=0] Leader __consumer_offsets-45 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,796] INFO [Broker id=0] Leader __consumer_offsets-15 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,799] INFO [Broker id=0] Leader __consumer_offsets-30 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,803] INFO [Broker id=0] Leader __consumer_offsets-0 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,807] INFO [Broker id=0] Leader __consumer_offsets-35 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,810] INFO [Broker id=0] Leader __consumer_offsets-5 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,814] INFO [Broker id=0] Leader __consumer_offsets-20 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,817] INFO [Broker id=0] Leader __consumer_offsets-27 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,821] INFO [Broker id=0] Leader __consumer_offsets-42 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,825] INFO [Broker id=0] Leader __consumer_offsets-12 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,829] INFO [Broker id=0] Leader __consumer_offsets-21 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,832] INFO [Broker id=0] Leader __consumer_offsets-36 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,836] INFO [Broker id=0] Leader __consumer_offsets-6 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,840] INFO [Broker id=0] Leader __consumer_offsets-43 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,844] INFO [Broker id=0] Leader __consumer_offsets-13 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,847] INFO [Broker id=0] Leader __consumer_offsets-28 with topic id Some(f9d_z6FzSde58txrT_Qj9w) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:16,855] INFO [Broker id=0] Finished LeaderAndIsr request in 282ms correlationId 1 from controller 0 for 50 partitions (state.change.logger) +[2023-11-06 13:27:16,861] INFO [Broker id=0] Add 50 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 2 (state.change.logger) +[2023-11-06 13:27:32,128] INFO [Controller id=0 epoch=1] Changed partition test-topic-0 state from NonExistentPartition to NewPartition with assigned replicas 0 (state.change.logger) +[2023-11-06 13:27:32,128] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) +[2023-11-06 13:27:32,128] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) +[2023-11-06 13:27:32,131] INFO [Controller id=0 epoch=1] Changed partition test-topic-0 from NewPartition to OnlinePartition with state LeaderAndIsr(leader=0, leaderEpoch=0, isrWithBrokerEpoch=List(BrokerState(brokerId=0, brokerEpoch=-1)), leaderRecoveryState=RECOVERED, partitionEpoch=0) (state.change.logger) +[2023-11-06 13:27:32,131] INFO [Controller id=0 epoch=1] Sending LeaderAndIsr request to broker 0 with 1 become-leader and 0 become-follower partitions (state.change.logger) +[2023-11-06 13:27:32,131] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet(0) for 1 partitions (state.change.logger) +[2023-11-06 13:27:32,132] INFO [Controller id=0 epoch=1] Sending UpdateMetadata request to brokers HashSet() for 0 partitions (state.change.logger) +[2023-11-06 13:27:32,132] INFO [Broker id=0] Handling LeaderAndIsr request correlationId 3 from controller 0 for 1 partitions (state.change.logger) +[2023-11-06 13:27:32,133] INFO [Broker id=0] Stopped fetchers as part of LeaderAndIsr request correlationId 3 from controller 0 epoch 1 as part of the become-leader transition for 1 partitions (state.change.logger) +[2023-11-06 13:27:32,136] INFO [Broker id=0] Leader test-topic-0 with topic id Some(Hx76FWANRJGp_-YQs8849Q) starts at leader epoch 0 from offset 0 with partition epoch 0, high watermark 0, ISR [0], adding replicas [] and removing replicas [] . Previous leader None and previous leader epoch was -1. (state.change.logger) +[2023-11-06 13:27:32,136] INFO [Broker id=0] Finished LeaderAndIsr request in 4ms correlationId 3 from controller 0 for 1 partitions (state.change.logger) +[2023-11-06 13:27:32,137] INFO [Broker id=0] Add 1 partitions and deleted 0 partitions from metadata cache in response to UpdateMetadata request sent by controller 0 epoch 1 with correlation id 4 (state.change.logger) diff --git a/logs/zookeeper-gc.log b/logs/zookeeper-gc.log index fc9ec45..b823933 100644 --- a/logs/zookeeper-gc.log +++ b/logs/zookeeper-gc.log @@ -1,34 +1,34 @@ -[2023-11-03T19:45:15.598-0400][gc] Using G1 -[2023-11-03T19:45:15.601-0400][gc,init] Version: 17.0.6+10 (release) -[2023-11-03T19:45:15.601-0400][gc,init] CPUs: 12 total, 12 available -[2023-11-03T19:45:15.601-0400][gc,init] Memory: 63941M -[2023-11-03T19:45:15.601-0400][gc,init] Large Page Support: Disabled -[2023-11-03T19:45:15.601-0400][gc,init] NUMA Support: Disabled -[2023-11-03T19:45:15.601-0400][gc,init] Compressed Oops: Enabled (32-bit) -[2023-11-03T19:45:15.601-0400][gc,init] Heap Region Size: 1M -[2023-11-03T19:45:15.601-0400][gc,init] Heap Min Capacity: 512M -[2023-11-03T19:45:15.601-0400][gc,init] Heap Initial Capacity: 512M -[2023-11-03T19:45:15.601-0400][gc,init] Heap Max Capacity: 512M -[2023-11-03T19:45:15.601-0400][gc,init] Pre-touch: Disabled -[2023-11-03T19:45:15.601-0400][gc,init] Parallel Workers: 10 -[2023-11-03T19:45:15.601-0400][gc,init] Concurrent Workers: 3 -[2023-11-03T19:45:15.601-0400][gc,init] Concurrent Refinement Workers: 10 -[2023-11-03T19:45:15.601-0400][gc,init] Periodic GC: Disabled -[2023-11-03T19:45:15.601-0400][gc,metaspace] CDS archive(s) mapped at: [0x0000000800000000-0x0000000800bd5000-0x0000000800bd5000), size 12406784, SharedBaseAddress: 0x0000000800000000, ArchiveRelocationMode: 0. -[2023-11-03T19:45:15.601-0400][gc,metaspace] Compressed class space mapped at: 0x0000000800c00000-0x0000000840c00000, reserved size: 1073741824 -[2023-11-03T19:45:15.601-0400][gc,metaspace] Narrow klass base: 0x0000000800000000, Narrow klass shift: 0, Narrow klass range: 0x100000000 -[2023-11-03T19:45:28.577-0400][gc,start ] GC(0) Pause Young (Normal) (G1 Evacuation Pause) -[2023-11-03T19:45:28.578-0400][gc,task ] GC(0) Using 10 workers of 10 for evacuation -[2023-11-03T19:45:28.587-0400][gc,phases ] GC(0) Pre Evacuate Collection Set: 0.1ms -[2023-11-03T19:45:28.587-0400][gc,phases ] GC(0) Merge Heap Roots: 0.1ms -[2023-11-03T19:45:28.587-0400][gc,phases ] GC(0) Evacuate Collection Set: 8.0ms -[2023-11-03T19:45:28.587-0400][gc,phases ] GC(0) Post Evacuate Collection Set: 0.3ms -[2023-11-03T19:45:28.587-0400][gc,phases ] GC(0) Other: 1.0ms -[2023-11-03T19:45:28.587-0400][gc,heap ] GC(0) Eden regions: 25->0(21) -[2023-11-03T19:45:28.587-0400][gc,heap ] GC(0) Survivor regions: 0->4(4) -[2023-11-03T19:45:28.587-0400][gc,heap ] GC(0) Old regions: 0->4 -[2023-11-03T19:45:28.587-0400][gc,heap ] GC(0) Archive regions: 2->2 -[2023-11-03T19:45:28.587-0400][gc,heap ] GC(0) Humongous regions: 0->0 -[2023-11-03T19:45:28.587-0400][gc,metaspace] GC(0) Metaspace: 8164K(8320K)->8164K(8320K) NonClass: 7262K(7360K)->7262K(7360K) Class: 902K(960K)->902K(960K) -[2023-11-03T19:45:28.587-0400][gc ] GC(0) Pause Young (Normal) (G1 Evacuation Pause) 25M->8M(512M) 9.440ms -[2023-11-03T19:45:28.587-0400][gc,cpu ] GC(0) User=0.00s Sys=0.08s Real=0.01s +[2023-11-06T13:26:23.854-0500][gc] Using G1 +[2023-11-06T13:26:23.857-0500][gc,init] Version: 17.0.8.1+1 (release) +[2023-11-06T13:26:23.857-0500][gc,init] CPUs: 12 total, 12 available +[2023-11-06T13:26:23.857-0500][gc,init] Memory: 63941M +[2023-11-06T13:26:23.857-0500][gc,init] Large Page Support: Disabled +[2023-11-06T13:26:23.857-0500][gc,init] NUMA Support: Disabled +[2023-11-06T13:26:23.857-0500][gc,init] Compressed Oops: Enabled (32-bit) +[2023-11-06T13:26:23.857-0500][gc,init] Heap Region Size: 1M +[2023-11-06T13:26:23.857-0500][gc,init] Heap Min Capacity: 512M +[2023-11-06T13:26:23.857-0500][gc,init] Heap Initial Capacity: 512M +[2023-11-06T13:26:23.857-0500][gc,init] Heap Max Capacity: 512M +[2023-11-06T13:26:23.857-0500][gc,init] Pre-touch: Disabled +[2023-11-06T13:26:23.857-0500][gc,init] Parallel Workers: 10 +[2023-11-06T13:26:23.857-0500][gc,init] Concurrent Workers: 3 +[2023-11-06T13:26:23.857-0500][gc,init] Concurrent Refinement Workers: 10 +[2023-11-06T13:26:23.857-0500][gc,init] Periodic GC: Disabled +[2023-11-06T13:26:23.861-0500][gc,metaspace] CDS archive(s) mapped at: [0x00007f767f000000-0x00007f767fbeb000-0x00007f767fbeb000), size 12496896, SharedBaseAddress: 0x00007f767f000000, ArchiveRelocationMode: 1. +[2023-11-06T13:26:23.861-0500][gc,metaspace] Compressed class space mapped at: 0x00007f7680000000-0x00007f76c0000000, reserved size: 1073741824 +[2023-11-06T13:26:23.861-0500][gc,metaspace] Narrow klass base: 0x00007f767f000000, Narrow klass shift: 0, Narrow klass range: 0x100000000 +[2023-11-06T13:26:31.373-0500][gc,start ] GC(0) Pause Young (Normal) (G1 Evacuation Pause) +[2023-11-06T13:26:31.374-0500][gc,task ] GC(0) Using 10 workers of 10 for evacuation +[2023-11-06T13:26:31.383-0500][gc,phases ] GC(0) Pre Evacuate Collection Set: 0.1ms +[2023-11-06T13:26:31.383-0500][gc,phases ] GC(0) Merge Heap Roots: 0.1ms +[2023-11-06T13:26:31.383-0500][gc,phases ] GC(0) Evacuate Collection Set: 8.2ms +[2023-11-06T13:26:31.383-0500][gc,phases ] GC(0) Post Evacuate Collection Set: 0.6ms +[2023-11-06T13:26:31.383-0500][gc,phases ] GC(0) Other: 1.0ms +[2023-11-06T13:26:31.383-0500][gc,heap ] GC(0) Eden regions: 25->0(21) +[2023-11-06T13:26:31.383-0500][gc,heap ] GC(0) Survivor regions: 0->4(4) +[2023-11-06T13:26:31.383-0500][gc,heap ] GC(0) Old regions: 0->4 +[2023-11-06T13:26:31.383-0500][gc,heap ] GC(0) Archive regions: 2->2 +[2023-11-06T13:26:31.383-0500][gc,heap ] GC(0) Humongous regions: 0->0 +[2023-11-06T13:26:31.383-0500][gc,metaspace] GC(0) Metaspace: 8086K(8320K)->8086K(8320K) NonClass: 7171K(7296K)->7171K(7296K) Class: 914K(1024K)->914K(1024K) +[2023-11-06T13:26:31.383-0500][gc ] GC(0) Pause Young (Normal) (G1 Evacuation Pause) 25M->8M(512M) 10.024ms +[2023-11-06T13:26:31.383-0500][gc,cpu ] GC(0) User=0.02s Sys=0.07s Real=0.01s