From 57612d52321c0878c66f2320ef0c7b40f546c551 Mon Sep 17 00:00:00 2001 From: zR <2448370773@qq.com> Date: Fri, 3 Nov 2023 18:56:27 +0800 Subject: [PATCH] =?UTF-8?q?=E4=BF=AE=E5=A4=8D=E4=B8=80=E4=BA=9Bbug=20(#196?= =?UTF-8?q?5)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit * 修改一些bug * 更新掉了一部分之前的bug --- README.md | 2 + configs/basic_config.py.example | 2 - img/qr_code_67.jpg | Bin 192980 -> 0 bytes knowledge_base/samples/content/test.txt | 1668 ++++++++--------- .../langchain-ChatGLM_closed.csv | 346 ++-- .../langchain-ChatGLM_closed.jsonl | 344 ++-- .../isssues_merge/langchain-ChatGLM_open.csv | 648 +++---- .../langchain-ChatGLM_open.jsonl | 646 +++---- requirements.txt | 6 +- requirements_api.txt | 6 +- requirements_lite.txt | 8 +- .../agent/tools/search_all_knowledge_more.py | 16 +- server/agent/tools_select.py | 4 + server/chat/agent_chat.py | 2 +- server/chat/chat.py | 4 +- server/chat/knowledge_base_chat.py | 2 +- server/db/repository/__init__.py | 3 + server/knowledge_base/kb_doc_api.py | 2 +- server/utils.py | 78 +- webui_pages/dialogue/dialogue.py | 2 +- 20 files changed, 1881 insertions(+), 1908 deletions(-) delete mode 100644 img/qr_code_67.jpg diff --git a/README.md b/README.md index fe062104..48946699 100644 --- a/README.md +++ b/README.md @@ -148,3 +148,5 @@ $ python startup.py -a ### 公众号 ![](img/official_wechat_mp_account.png) 🎉 Langchain-Chatchat 项目官方公众号,欢迎扫码关注。 + +[Langchain-Chatchat](https://github.com/chatchat-space/Langchain-Chatchat):基于 ChatGLM 等大语言模型与 Langchain 等应用框架实现,开源、可离线部署的检索增强生成(RAG)大模型知识库项目 \ No newline at end of file diff --git a/configs/basic_config.py.example b/configs/basic_config.py.example index 3540872c..6bd8c8d2 100644 --- a/configs/basic_config.py.example +++ b/configs/basic_config.py.example @@ -6,8 +6,6 @@ import langchain log_verbose = False langchain.verbose = False -# 是否保存聊天记录 -SAVE_CHAT_HISTORY = False # 通常情况下不需要更改以下内容 diff --git a/img/qr_code_67.jpg b/img/qr_code_67.jpg deleted file mode 100644 index 719207f93309cec5b8a75d0bdbe417278ca6b39c..0000000000000000000000000000000000000000 GIT binary patch literal 0 HcmV?d00001 literal 192980 zcmd432UJr}*DriRfY78#l`bGi?;S#sE`lf>3{{$R1yoD`rFRfeIszinq=O&<>7pV< zs&r6kngGTKNw|mqQ{Q(z?^^f!)?Ihq!9#Y2J#(h)y=U*=%y=?;vH~!g7#JG>5C{Mm zfj_{>DkMTb-1{Z~n3(}$002+{ln`bB3Q`d82Y{RdDE^`Wz#78!Px=N#^zS@mAdegX z{svgU=VqKYK>m0Bd*J730^f34)89VjWaQ-4WM$N3RruxP)D)D}WMlz=AsqlSfsWkCK`4}=T~WTyX>Pc9=P17v3W z8=d(ty5t}4m6eexf&8sUCno?dk-36PHm@1DgT- z_s^3(Ai}~c;C4WWS3n@Yyp#-}qHAnMema5xaETjY4?wwu4Y|S5B?HK~$>0E`oScHP zj4U^q7C?3SB3T((6>c)+fB&KZ(Eg1He3^{QU-b$9RUaeZvL)3=Cg8;6R(%Ahz1uoj zBNH{&(?eKTnj08j*8fW=$T(aA18j42E_gx?uEawS$0k^?2I$+Le4IUtE4EO_XfD|AB(sF#G7cd(<>mhRJA_Ub04SDDP7W@B>W~Eh#PpMs<7X!)#4-SYegS|U|C1hoiHfYB ztRD(F50EiIpiGdHet-`gDGJD6>92~wU!dZVQ&7UFsA*_Hh89p|Ay6n8Ih2CpRHz{j zP8E}!iGo=`R)_NJRTtR#AQri(j8ZB=-G*LP%W;g5ylZeYH4Pg(2j@BA3nHQy#S|2k zlvPyK^e*We7#bOyT)S>%Z3A}K&E4asrUsSzpY-(<4ZENp%{kE@vU~p)7p%3ESsU4zGI*n|{Tg|Vky`>SXFdmW4Z|JAd9cI;pMS^)F_vcD3P zj0{Qvg+eJPDZoTYMR}U2XsG^5G=EQYe!&(o)kh{(nv< zi=aW3JeddRpb$`)piBT9AQGM>iUR+SJ`foSNfrx;QSe2t60$_m|MD=LeENU7?f=u5 zLQls3*=}eGb|_7Q6*HrqKLBS`T#%;!j$yTa-CZcTMz?%NPZenB?co%8Ri9i;qKP@C zGZ^4nvk@O6M70E~9G?G(hM`uxb@O)?7#Q zkCm61oE6y|)qMG*^D4d2gNp882aYiX2=P9uZye@&wQq zKLJFK&?kVv%sx|NO3A%XuwOq<00wT zT-ByvvDuFUn?y-0?N-}DwG5>l@-VDXc9whJlk8^;&B;0U2xp;BvoMHfTIljIzRQuL z>S``h0`Ps_WX0b-0muXJy;!|9MRY$AI@B#X6=CdM#v`@%$w!&tj*y%au|jx3sG=P5 z$aMe5Jk>CeY5s%H1Q30e2tM61CqQ&`_`v|`4BVX|zDNdRP@$Ak>!9O7#~!*NF;3gL zbKjdriP~CrxJRZ>D!5#R9C%Qyk7AfcH_zi5$W~xB-Bd`RLY1{+Q_gKIJ^jR#( zm^HnXr9!{u5mvw^J&(}G4)vMOxjt~{QTzPy;e!`DY}WFQbHgqL4knFViJ4NI${any zR4VH#7ts0O)Smz?411i<@H2W-yCh+JqOo7yiihzp)BJeEH|VYOQ9*-kLkc@1!DvzF zvGZ`5*gfu<0|9W%&GEmm#eKysl4B1eB-c~iE{!Yr$&Kz#zkaYgeKoE<*;G&_D!o&l zY1DlMbgCiSD2(A)3W}{O09!q%bp~6R^vj%a3)=Sq`O3$}^2V+9=pRc$;qx~`%`KnA?_(3>IGb`H!Xd&O zEBE2nj@a3n8#459fx`CImXING#(PYOq)9D~9ksZ45*OaQU#+J5!H=qPTc<0>R?4HA zuhm;WLkw)QA^{$NEK>co=(9)~|4k9#08xOzTt~+*so>p?#cud4L8j|&^B7uuuDQkg z^~3vX@xsjVFOAy0MsER+`PT(c=PeCJnnE?-j9HiDvu?rXuc-<*cSdVBZRaX8$=bkt z2jshgFMhXrT*b|l3QpV2XX$iw_iQBBHFb9kFvyPT*^$jJwzFTxxSW-EMW0TsY;ZHG z@PFGH0PbA>QeUrw=-d0bCx5KJyfpUWON0ehwrJa?$Sm*8JlT&hS(-+%7ebhq05nn- z@dADB1n^2rg_5WXG1mszx)r8sTk08i*mKUsmYwO-xgQkZqPP(#IL0e1c+0q>%|95e zs;UZ*(i-7n_9Qj*O@?{BbppIOt7Pm1M-bB z?!l^@FmlEoCSp82i@)XzGW*l9bH44*^6*d*p8!f@8mP#otDh_15}`f3B2yZQo;9tt zL7}&Ia+@402tG`ug8hoqzf_*iP#baoI0C;ufS`xFo5y!cVXP~Z z(rV`I=Z7f1kW;&|E5Fha&_!;YYJIQc2BWqD1gFVL(6UEP7B}BX`S511en6Uq+HXzIw?v`_sJ~B1pAcLM`O6PmdHJw)RnHc-i|AiX*Uen%xX-6W z+HKGAR2r1)c?7W_Bz*kL365ok-o@`BSH^ruF6fn<6M(l{0Uw8}pXlOS3*h!u` z^?N78_r=oaeJ-wWUawi|Ih@?$F%T)?hmcL|l zCNA}Yhl&O_(ao7hz3l1g5UhItvR7C3X8QB`uyZ-- zIVzVjtPIQ{+Rzff+2)riXvlwJ$C9)Jkwe7{glp@5S01=oXR0rFzHWCuFy<-T#bJI% zknIYlTyOGtlH4GasjanG&x9X3)+aFIA(e(N2u|1sxpCIcwJgkRIX|p!&YlyuWi#He z2$#r6@5>T3>f4=2FuVL)(}fwP2 z8h|{(z8mC7m7~;0u{}}rng6=>G5n{bu$Pp8$L-s~PCQqv9;fP%yE6@gz7Rf8<*=eY z3cXSQZ)Q(`dlk>+n`8NsD>HxhYw^4()6hUa$!9fBzI*Wf6Lsu*K-Yj1rj;=vkcowdSQW9E57u6F?4aLUCMAQ0@w71DoQ!5s*HPD4qm0p1p8IZz+UC>Uz;u%f=z2yXyI-F!!M&i!>N3AHZ*5>O^Ss?f~|DCN~ z3jYCA7dK2%3})SKrkgo@?AjabgQo}sHQ~;F}qTRa|Ujq(9K|zhkXG3Ch?#y zWZ)a=qNXq3!FtFE@N7O39_>TS1tm`eHPVM)APsoK4_{BC&zaj3K@0R>b0gu;?M2Il zXaIWxjM%Ehj_OryfzR{Zm?B9Ku1#xPTs!_8?PNJ*(w0eLDR#E_IO-d;X*B-wETc=* z!)E{bWxt*R&pdFcSSddjT3Sh(M=;Nm<{ez_gav3^xPI}g?sNl1h8*Rl5Vy?A%6bKW z+&NXB2KWc?*s@&i>38s%%ADK%ae7yy21}EM%3V6Wy<+75KNu+2Q)e;6u z{Q9wJ1Pd|$aN*3gb-sjWpUOVFd({dunTBW#P&(Ey3T3i5Mt*4X$t+2?g??S8Mg{QA zIjn~z?{)J^s-AF5)bl7#$QD^yfafruDp<@_2(c ze4*zjZfD5baP}ipvZ+Mu)H8Od|5uuRvS4i$4oWn*W~}7roQpXD_9`*6vkRkvl6C&i z9}gxkIhuOQNa0~V(b_^Q?y2`IV21%%S`rQ4u>e5|8(X;|%@MxkI(^$aBI!kKO+Wzi zMQw`oQ9rtjT)q9q*mtkQwD}XaxonBWr~q_i9_T1)$k!`%tV(+2b5%*w<+O#)Fsgmv zlzmq%XCn^CeHVN`Kr>`0ZC53r`rEQ!`;{e~0)E6pYikQlJ^rK=Cz)~N0;ig~sbEO`H( zvG|iwpfpn27QO*sJ2c3!}1)#lN&r&Jsk zcG1Zz%-Y;*X!Qu+3ZFc~Da1*#S)siCRRNF##{^~N|ouMwn`AGPy6Mzek zSP>zx@ZpeZH=ovNRj)c-xm`35##|w+A1XFmgPv?UmM3yo8_Gdjnxe9O{=nWYV+@Mt zmiv|jC-3Zxv4p2t#hDk1_!@h(rmEt&cW)+A1eQk?*3hev->K^(QBzdj+^hYR57c0$DtiMaeI26djPLJ22mH!WeWC!LE+<0jW6^;k{=5 zs-or7XQn-m=s{zv*#W3dz#IEZX%YDj-ZULf{e15k@Zvd7Uo0SJ?^Wy#ls7os*i!TD z6&EDWw&Y26bCv0}WfkJ3;y<_nE}i688_}qFhQawvhj+JJFkKnOGJ2a5jR^Z@ud$TU zr&r%a2kSO)L$&!PskUxtV$0Us-WJatyiHn1B;_66qt6u@2{_A=>Ftqy0>p~C*klaz z$@^W4AuoX3=tCsKG4`Ejbk|o9(bBb&6*<#m{jjlYraYC0^TSnd;%M73at7vPky<_Y zQ(b2wW*|kwhc09I9%%#%O}8XmVoU!ryEHGQ$yxQJto?i)>Ht6lunOzrx%YJ5EL zar2&`HSu{ji``c1{-W6lKy8pFU}Vu5zt)**%tb>zA7An0(2BeOFhlwGIJ-u&fxry| z48MA8OkgJ#+6Cbv&|7NB>V}5DrzIrDdKFaBl^1~&?4!7|L76hcLDE>)1HCmH*s6Fud zFDsgVx&YH9IYC?;f;osr1s}&VowI`JJ6usuH)R0U~@kF8cw-P_MIfOY2@1!%qo zPR&=349wTy-Mf3N{K?U09(*{~2uIn2D{-^~X(mGRNso@o+ZO@f$25L}Zqe0$W^&w77R>So*1+S|uF?Hc{7(Prcf+>yL;H9CUP z`VUqNkl6&T@ZXoG6au)zADVow6>jWmMt7vziWsEYE_ygyjYoR$V{|xKAVhc4xRxX_ z1MNLF$9p`erHXyoi4;J+zIBf?qmn#Zrc1)4Rd<~ZexAe~!2yow zbti%j=1?}m7fV?oxa{z#rMoe2a3XX`X-ILTG3i}ODpdK&6IkqxGP32qdk3gqfgk?t zX0K(X(B~M7-NOuFh4(Jc=i`EjRg;dG`nM;*S;P7_Ee-6`!8#!x@VT434k~u*33FP` z6~++N1$ zn`sfzUMNnV1V2H$fmLtUyKiL^gmaF@GdM}#Z!&dQX{47&MFv7};Ob9cLg0`c==Rr5 z*t7&c$<2&Bl}kCvf|kKAIBWNHQz068hXL~sasI^gZcfth5)0l8o7De%i3L;qD1f#3 zjjo9NHuLh831d~pn&7+71LO?u*y2hyC}=2*OXS1}KvSzw?d@L&OH{afIj}T?(C)vUY5g`j zwA|wZ0)7To=G~%US z##~>XjS})2ORM;EMNT2bDEa|}ZB|q55JV=9!6gRwNpbNSE{IsOaCx#Wfa2$)ff`B< zr~alS-wD?=FZFWB<2_D(rlIp_5tzv4w=HeHS@fR3@kP)|zmDKrJria*Q)fA2&Jb3y z{aOL5_Q5YA;f=wsOeJHD_E~OsR)CwCAw2>L4!Io|%bvj+8sTdzeUGL)s=j+HP0$WS z@f#PKP*X(<-4v^cB;OkufdMv?%MH(qUwFeCbKj<@UvJI%_=ADLd(G&k9(R01iMEF` zm5N|A#RJu%AZogUDP1rQ4tG%3p08He|7!NlkZFXF|JP*IuftTz8>X{?+F|^ zFAfBkM{=+V@m()dDPb5CN8rzo?Gh>8%Prw2`*WGfs*M?a_32AC>h#nD9l zGz^$r9Ct@MMhL}A2tci98k$U^nr^o@_XF@blrlB|KLjV+ZgendWe&eTV*F0|@x)M=x=|v2-p=Y%I=z9829q|Q3FtLyk=Q|( zLu7D7@n;kta|@g^eEQv6RO7tQs}Wbhgu#)G7~c6WmshDN^prjew_>Oc5uo+y<%WBr zS@6|XSmELXf=<~$m7{TBYoM{8kPN5s%}a7U6RhNf&L`|-Iym;Sek(}9_P#$X5h?yTsxQT?e4**WV2?hP~$7q+G+zEsq;Rhk958U(RG z-&1^fp?+Om^h@{%w2GC`j7MCVQcPpox@VmgAi4Q#af(Tgj>YCW}xMj70|M3Ps<{_i^$vl0+~CR0on zu9=vJ_71y&WGMdMx=?cTn*QR4wpPq>+%#!8cGl!zKf2 zP5|GX(UCp(*!kZEllIg+G()1Paw(e)9L_aVDk|j82TpthX1MX^0Hk=tm5!#yyfvm3 zC*9eupeE^J%?}NU5|IXflp3tIbqk;mwdrP%XHdN&dy0z&^U|qW7ZO}MvRNf}g*u{u zV6q{Z$yt+Wv+!Y?us0Rkz9wc8!!aOL5hCOdy4 z)UUonx1;Cv(&m=e(uj$#mr3cPh`uMs%@EwrUFBn51Zn03V5&b9PqF=uJcrqpb=+>d zDtbS&N!E5XiYsN`FQ72$I*l=1VC~T_F!aCj7eX+WfZhjjrevEaP_!A~8(#Rf6X1qD zoRs{!{`zTd?vlZ;W=#Q*-CSoRTg9tXzo!2zdOL}xn1Wbdww4ZkL}p;oGw_dR%}PJsGFFuF2zt=~wDgjbH?&({<3W)e}m!64=_ z8Cr;V#Us8CB~jI$02i9pijTQ}q7N4;ZFYzN%T@iq9@NR?zhOm*#OgGHhxDLkr8ln} zmYPT12UMlveer4dO2K--UIXS$_z94<(EW%B#8{Pi3v4ow(@>J!zk^FVD9jbaA5B!u z@~`8t6QJA%)Sc9{*KisF8K{+`1Xp~@Dw-9~x0)7rXS7gUyKAz1>PJcQ(-md;KVKCN z5*eM`m^N839GN--p(Ik}7eq<@*FF-h-w7bx*29OBhLf^KwfWL(!AO-ms~B(8!n7;( z6;>k{#wot2D*g^P)>l6wj6+jXv zkgSHTYV*xV6PSOm#ib&}Nbl_ZtJjH^PCh4q@s!lz?2jMR_Y)*nX2%oCT`wn>;$k(j z0cIm1;=53BGsJt8;`4isG;^wIc8=0!pQ*(?nZ`Ub za1WL3AMebh1eir(S?MxJQvJtzyy)-UDsXV-%T{kZd={bH(VS+3!BygVDb_hM!#9)P z_865$XTw6NHpt`)$RK}y9c!U_owQUUwDIFBV?g-M+t<)7UXjm|3RU#QtRGtX%^ryL za=+aFa=F~#J?G=f_Z>i0Z~bziX`dF?bfxJc!lb3Ukll4efm$$TL|w=Eik$J#<(Y%p zV|7r^E~34QDIKQG!YOF61=1kUW3~lov z-E#-W^b>NjmVYQVLU0pW`0V2eAVhC8w($BWC~6QiZ6WJJebr-0aLO)U%-EIA)VDkR z^JYuMXZl~-t1N}jU4M1MWFvXKO+{1uPXiyuh+zSscm8Bf>^}hp#Q@@2VjXIw^u|TJ zAI5NHCPhPjrB=}5HNK*f=Z5W_?E4G8iix5egBfMIPDvk?hHZhD1`}5dlSl*LdWDKd zvw&Mes524L*s!QB*~(db@mMpvnq+yajf|0?yg2j0<$BJ2%t*|Lz{4}`T4l!$s9trX zY{XUkMT}n0GQ%%Pr4zskk0|fXxYtHMJ))AM0?A_+Xt#dIWlN?9-G*f$2N9Ud_4Bl= z40}wg@Wf)K%86zoU!Sz*XwvVZ-5W zTbS%ZsS5|pwX1Z_#@i*G)A);4j#MJ$dfH_yLLNB_r}%}Qr^N+hbKj!QIQ4nF^tI^0 zO_&!8jd+K!z8ma!f06ZWt`p`T5=}`3TXiZXu;t>?Q%SI?E|FEcTbcUu`gH8$M&GHv zFdK9RmM2G=$iOnG*3e~U9N&wXx&%w{X_gbKnUYsNF5y)CnMC%x$Ip$l3AynzQUlP` zY{JMaD0ciPU)7)dZJlbY&CEvCLX{Sh4b(vEIHzRPN0?@uA`y-k*U@6ZgvP9f7C092 zzMX#mj@|Q#-ZXmu-+guwCo-hsI>VR8Y-nTK=2Q1rKqd!vZ-CDDCiP0_hB6}It1JD7$t{P2~ zC1`2l&~3j7Eb>Fe5Tr8SBN8LVGp;3Ep;}bx!t3e1@~*JVy|5_eA&K*fRseU7#NASgBf9P||WLP{Mp^O(5D9lAmcg__vizLC{X^;lAE zxP3scl<;A$)Ti<_vk#3R#R$^vlTy2POliG5V)sLQg_!60lau$Hw?|{ioxky?LbSlx ziF}FpkR*#71(#NWDVBDvVz7{Ac}v8kwWYD(OLMHb&zDCK*Bn^WtPFD|6fVg(X&7A}MryP8JM!fh04VJ? zDfx^<0vO;#ryxgpHHj}tEMDheeZHfjEl#lO?9ar9WDPYeFMWDtcU{KXsW(!PPGwky z9D3k@yA75pk1)k555dXta8NxN2*SS4@wwMqCfsvP8uKQ82);0B-8yeF`L6S7fxOh% zQyq)z1!>>E-#)DWipJdoi(8Q zZ;cj*Fp9kz#AXR9Z0qGhbTgVWX$TLG*}-xqT0(?t#L@^w%;-Q7C8oRBWi3Rpa{ji1 z>a9BGJ1Ng?Bn~d~z{8@SSQ?WnbHMFTy`a0+0D^Eq;IXCOXmAckUdM zkj+~%9wC}OQRiMu)y4Gi-ajy$0@yQm(x(%E3guN#O}L6rP|9*C3->p!D{ye)dRh^3 z)}Z0#BMUi%a6%7&&7=T=i<`p;tbA*pkF?anmwSuOVPKEGy}~S2hg29UReMNoE)b&) zq6#;R*Bs>qJS@MF9zu>(HyeJ}3!|<4ea2dZ-i?Wl;WR@nl<_=Da}+fsVBRx38!DYMVHaZ*)s;Z95j<8D&n!D zt-j-NX6rqS`^5>^7p(_X!Ke~�XmaZsD}y+G{(23Qj=VK-ES^z zCYUvw@?tvtI$KXINtQN`>JH+)O+NxJZaAPy2&$_~!0hyvT}j2cw~{`+D%$=zY9LB? zAldILm`-W(ENYUYerqhaLA+r7RsC1!;4;Q2U_R5FeJ3;g=K7_n!?v!cd;T{t3tbHg z$?raJT8qxAi01$p>On8)kDUO#Xjme{E`!J(1=z3~Ss)-vZjUAiQuf2n8Rt4@jqS^S zy?evKPv+0l#R4Uj`k1b>VBq0{@9>@8Fd2;`O94coXWO77M=d2kje#LE=}_%C@o0%o|@uZ;w~O z!wo#bY7ZEE_6oLZnYQ3OclmU|&6_;BpX#iO*C2-O;s zrX4k>=cPh!J}v#!7p)E{din3)#?{5E6+t);S&jhgaw78Lq!x_8i%sj7JZl*94Ac7b z{!3yaqvV8V>!Xq8#~hXm+9I4oJ|zDQupiBYTXM|B^j%~&@QZ5tBhGT_#8;s>o^~Gb zBXiJtaxEto!y^1JGiI+%x$0TGe$U^Me@AP`l55fI^ByLw!=J8xyZ&n-7}}>DA|9 zbX~R;&;I)EN*=T~lL!z+h%dmMOf5CMAx3ZITR$AYU%O(4Sj#Cm%91MI|FU2=KprUH zUroEv)TFQzDYQ!DS@gV*F8$&iQw7O90_))7o!vrHBOAw+QG8&owjjG?Hzr$vZ45aLkW&teCZ3 z?(GIM*_(EY8i+vCpiG1f4s=|jZz4~vDd@WeP!r7Yaavh!W#O)N{<98_ic?L-o6Ww= zC6}P;`?duD19>_X?C?CjegyIOT7*Xzu^0)2_pRabE1y?~t?q>9(Rw?Un@bM#hkkcV z4XrlfjO0(JiKeD}Ly`+U(+A^Gv-w!uQrY{)cm4P7&vo-hNUpAj=ARXhz4V*g`2T46 zwe;JF)wiYBxl-9en|Wq$`SJ5TVk^ej!{<%Z+ix7aT{@dH=X5z^$S)#wm`Z|m9vj`j z31mAn2pe(KH=O|fV|Y>NB+?m%xN|nqU0lz$&DR!b7W#v`J6Uow26&dNRRmKn3iMT3 z+Kn=XLk#3lS~aMZbiS5117Bnu>*D-qUsu1Ij}l->r)9NGVuYzcwFQTmEZXlT$B>Ndj?cK8tn2A zLuycj2sZa&glf57?cA49is=|{#m7mC<$khfB{Dw>SzGh~*Q#H0-P{?2AQWX%e%sP$b#D5pm{5LFjtg zyU6qHt!Rm@h9J55jo}R_h2DeysO3edhdy7_B$td8}PI#{5m;YvCToLixg7pT&Ao z+W8-*8!Ce%atd08Mg$W}wKVt=muc&6k9M9hcssx3CG}0IabeV_8(~>fS0|&xAxO?B zLp?`~gs&8#IYAK{8!VEo*9J#(Nb8=B4# z*AjpiFCL^N+8^zYt)ioum%grU>S*O_9FX4>zSKypAkPF09K|*^_BeZpIAR7#X%Rep z)%5cbQhZzOy1nLfEgg2$*B=#_IX5O;{iJu(?Ph#qmC(!dihNG^*itE|SoerEdo8s2 z_hFx;_MT-=UyL~x$%{;{Zyw100;p&tugv^h-u1*xcUiSr$q zTeW?6O(i$qx0#)5P<7$Vy}1j_4g*2jJVSfudq^LbSsX|eW+04k3JQICFi%=Td12G$ z7S(TfGCuQ2QoZlqHxRw$rr(s*+X5S!9F z4ep~tP*|%&*dFL9i`>Qfk_Ne!iS*r;CG>5ivuG~Vr@j}U;d5LjSn**B!E+(8JcfTq zAF4B-MQTd-!P!X{!}Xk+u+p&vr-GhGJV!X!)o&#bu&><}%+3^o8 zumOGMRQP)91>Bye(W^J@`-4*l{(cjN^Xg~b2ddj(P>4n49@HNJ&=-d9I!(Y2Q^f%A zoirf)bHRLV`G_!qI&2jq8C4%Gf`V3$8p*=9kWz<&NmHL^Zzn}O`1Wu;@4KRu&)}S`FkTM`LyZP7g zpc^fqKx~3|-VEO91VFT*j`V|Wm>%1n03#Cw7w|2k{qUopqBp)I>eKP7$8?g={@{cq zu0_HdrjC;iSqU@XtoflgS_yEF9h}v}I(yRNvp=Li*&KisP_%-$+rkRNF?;a|V6q8f zZ6>G3Ai)RzG5{|)M#uxPsfIWZD;^J=*~(2qZ(N%`0e0T)fCtA~kT=1)c~Kjm#)!#k zX{4lB^ko9r%=1NUAQor?Kf3q=iWd#&^2OPZs0zX6F{}q1vzDI#6Fc?ke+%y z>%L7ahkGx_EH{ollmrj=YMGV=pl-M3v`)(;{P297@}a>l1JI{<()6OIXOyuV`u#OY z3p_BdC4`kKJH8olsd=hJJpakz=o3jl)(yW`4oTq~w(Fd4UwFDjMP-^>`u+}CMG;ux zI9GfX7OIc;E5FcEv5n+!H#Id5*gJ1}v*U40tWx;R?8n0!=L?^ZC#S3O3=N|3BGS7f z*l%kTzx&5Hy>qAMzdp}Ms+|1sp(~`z%`1D-mqF$QjZu$)j-aX#^xy;_QDv_o7`J63 ztTDtNWg16LkrPq^Eq6JeCHkhdNTkMxTL&@!5UiuT7j!(YW;4P^4fkB0+rfdsLsEqC zs*&eJbFsv>+F;pT!No5RiashiQ(JwplyQ71_AP{$T5R(Ol^3|XaJ+{ zIbi-s_Z`3UBP`=4<0D+96m}fk)nsy1y@&Nv%XqP|WP(Fgc<#z{VU^E}yo2eA znflf6JYB}6MPsoKs#k}e@~7x2Te8qUWIf9F1obbOn7SwJSI=cWEoHrAmEI>+K6p_# z-9+U&%-0R(4dbDzk06%7UC?tsaT)pA9r&oSv4D`(PQ~TskpI}}S2Ec=^EXI^%$IkoA<5}+(5xW^$R8chcKBi8 z@t(wb#WQT*EZcL^U>5Xibs5tG7GODV443wQk^;5AGYWtU)nq>-k%V} zR?O5HaORN(zsj*n=Rx@;Z>4W8mtS3msZfbGfV_9UAbVdA#do=tlRoB!zA8H3F-f*8 z9sikk^{_QvqKZ1D2rm7>tG{IR$&;f<83)dVJ}W$pA;F$Svo=ez9*f)Ixxq@mEmIx}eAG z;5onC2zkuRx>WW9_UJ>?cIQuf%>}K`rOlu0Yxuu^X|FHWHE2LHCls-%51`bS8>N%F z%u7j=;P!M~^8@Rz6UGG7^7&!c?DRKpK1>CE;BsCm&*&6{_FC$!pC(XM@6=)t zY@HjzJ;2AHFhc`N?;Kv}!#(@{d)aN-if^h?HqJt5P3p=G`5_$L*Q3kCi}M^v zv;xN>)Dg{-T~~I}m1$|6;TJk#hb0Wt^v&3wp^6AkY#MHtSgb{jhxLM3T9s`^&5@))MDw)yK@rKW~i3e$k#o*oSc`}+rbsu|mdRfIwuj@tIrk42H-&AdE4Fl+=ZCzVH+{qDj2{wqM6!JXS<|_a(UsX6XL%kWH}^WT~_+8@zsV5KcfMa zpkTWAF9eAuf8CtJXTGiX0o-a@s!sA7=0fg^k+Vj+ihHBOQX*n9%EQbuL)?43*;&tg zK|sFZbB)$yIYE2hI((G)qQ?&x`|Ts+=vNNx?`X1xkf=N&EhsOM;N2SY<4_C@A-Lgl zT|L-rrbCz27~4%J&SA!YgL5if$+7#nKk1YJ@@U8(aEFK-3}oVK(Y|j4FEvg{@lUom zsau$~lc#Jt3dtyB#911tyF(tY^QY2Nv;O&w% z?sC0%$q_mj0z$dzXo#T086DvRh629zDIp64=p;eUb3JZy^H?xE$@d}9Tj0P^_{q5W z%)qM!IhOT15l^u;4qy?Jw1gWM?j8nketGwMGS`pgZJtYR(7~BkdG0|>xArKsN#7U1 zuvrx3)W^p;VPU(E+3lx1r`Szr0&4a&e#>$A<%sgQn9aQv?_}(P6sfGCEy-T|BKjU6 zR^o<0*K8e!B&PalU5t=BzFCOb|A;dUjWxRbx@NwnJRs3Zgdxy?er=Tu5=%v?ie^Cd z-axzWp+J)wBVWTb2OTt1Rv>S#mbPFP4KUZkHL zcjePqK;@z_*K<8_q085+xTC3tf7fTAyg*1wxrbq{axk5sh3);Uc*bj?;{B_woQC(3 znK#^+yBgn?7Pv=t#zQ5OlNaz5$04M#;ueK&Va)QQyrT&}cJuoS3r;uY&L2p?KaYz2 z2{x&Tw-t}p;ff>Y{sh~Hyzo6*s;F|@4*UI^ojuI5YhqNWRw&bCucL6bpy}+0Nao%8 z>Ecw97WVT`0`t&{@6t|N{e`kxgq>qfVQ=32Ti?%Zu~&55<1gj2wmVfv?I#Bx{Q&JV zd9vExu}&HO&=6N;k|anozqb^$LH(SWsv#vGvk=to{{fjC;W~wO1~CE`KEQNMT+<*R zX3`{S^?UO#kIymNUB%vF?gE!e5=fsMcSS;}U7tt7VJHu2*6@05;&Ux_C$HDF9}4#I zSthMF6RmYIUltAN#tXPvGnG7?!^Mm=&XHzWHs}iSNWtivDV`*1FD>@ zeX!Jn@JmPcmKz2dGW<4fE`H5s%B;{^X$6Lf@_5kGr4S_+z&(G)=b~u|4#b8kUp4zn zpqJewDVW0{lpeq&KdKN{yU^0$0%LAh{a!M z1p|Bl9=t^5{vl-(#3L?a0}{H7v9$fKf5wCNgZ2!LEjsRgD>AstLl`8_H0O;B1Xci* z`t4&J7`8^olEzTYZK=zvZBbXMDt){4zUZ)3`f*$nD% zF(#-v_Cav(RpVMv#rC~hubcM*aX95Av*s(`Bm|Pn{49e*H(q>X0sk)!dN&d^Eol4^ z#H<^V0T%fdPSXAC)1lqBW6jSm^0%g6%1;oBD!)mi!j&qc!dJCkQqO}Iz^Jb-?;fns zB4U^66mYM`_sTOv54cAZ4n|WB%mg+IBcUpQD0tN!IvZyKkZiyxwysDN%d*^3JD3pi zVoa*AVoQn7z0I3zB+`uc56uN_iiI40_Z4>xaEQX6{~1rv1#bYuk_Z}iXdhOXu0E~v zXpw7(`Ifl<`Fwh*df2n!ZLSw|+ADd)Bx1=4;Cvc8o;P4{!M#J&i2vzo?Jz{^HLasss&Y-v}vuY+wjE3dZr{xmH953mDerI2ep*mm=)JSTpuF z2n^IeuHIf=$Rb@iSX(A0E2EEiA}aoNVH0(G83JBbiX&)%ykSOQFsV|Fw;Q@g+Uf?u zpAm-BE4Z#dGQg$;S-?QJ-|FWdITZN5nXh0=D;jL}^%2uQ39R=?mx~9`zy5&H=+-SO zLI!w~0gTtlL02iu4r@+K4O>3)o2>ptLi5A_LrDJ#0uIUP13UTre<*x^2VYktv7=`8 zmPr|o7BHMzaS$>ZT|9=DSnmCqz$<_ah$4yL!*pMNCTMy;AAP=P?tS*7U**}SXFp!I zyy{L*CcE;Ae`SU5JKckEyd~3&Apa*A#tlSWMN;A4Z-56hm;*=`((pPFK_`GN4#Zs| zvrt$?;V<&>;&i^2TM!;TWRzgItHD~Vw!Lh;PfI< z^ow{ZHgxO(Ar2vW6$zjWm-eJCb_)e0( zrZ1_>A9Z-mMq5VR*+zLShhde+BTmmFWPxx00~mpd|7TQy;N#zNEDp}k6>xsyux-S| z3Dnk|S^#yq1^*4S&)FEyq=>5+o8G***u$hk%LSw9CDT5?vzJ|rOJvJbENm|ZD^k-LN9ph?OMPy5iMB13lk0C$S$ll^Vr{{CtD6%mw9NSY1!xOb~LWR~`(?RL9{i66D|Zm`oW1 z`LwN4&tpe;}yt= zCipgXy&f&~P#O)@DmG1^DPLrP%lCNr^xmG<17fy@Y8_q==ZAM$Ek0{I`h0udCQf1J zezkY*tyhEcRkuSJ|VNh zMx5wGfoFM+0T;PhRlRW~qOqK-!h!Jl*Tzkjo^#6v5x4E)P8)Yz8(U5@V}^w=IQxlv z{TYm@xwwR@MEd5z`P!G>TV6rjaFvVfvQDd<1MFW@*k$sQVv>YwE`pz_Nnz`I@71>? zfoZlM+fix27#Mu<{E2Z>YM;t|U;N?z(bpGjv<1uuj%suQrOloAj($8kb&2S-!%E2mQx}+l zYy~F7yk$y$Txa}`uP@iRtBKB`@{R$KSz!3L8T^kA`t1VdwBI<*zhP`gSR-s9vQ1Vu8=W5qdJ?7$px3N~V_!6Bmyqis@8&a*S^}C#M@U z`Pq)@UFhM36v4m6k!>knB=}_@5D`1`lHy}mxY;g0UcJ`)ChB@qLu+E_ z>UAT|R}xx+eH#6GH2?>UP7Tu`{Vv=JU)i#>;k|{p@#Sz+YyC6dRm`@`aZ>+x`3l)b zl-ctqg=Lnbl@At_WhI-x(NA7Hi`m#(DuZCP3*U-%p%PA}Zq8F}oxAamcNgV}tM~8a zzGO3P{gAS_o-39Wsj@PX3JJM3Z4hCNi)_1-+(^@}DE@Uk6SnV*h{fIpyt)XPC7J?7 zeZ@cFz5N@<=611hbsct@r*9K2RExcZ(bw3#Wnc|4DG#N6_S171oGnyuP}-$8bwW0^ zD>V)CyKchfaQ>+eQGzOeAs&ArvjHRcbQ+8hmZYnFPgS=SMsyi%>B(E}sjEFwFu zzLLV!6kwO1V|m(ThN;)XPECq--;bX6))ib>7*uFTkmOTsi+5dk!gAr`Ys~;F=wuYS zw6GDdSlCuVx>KVNFauvANk!Y*hj1_ZppiykX>t0<@0`i4@k#8c)@Z8=7o-R_s6znB z&xA_UA9XWWQa_v7#pf+gl^I zP`pW8mh`fSBV{{}ev7rw5?md~FtRs$#@y~ti{*547h%kEC%%)?0jA1ZxY=pEdQW8y z<*KtC>*`wzPGfk0PQX87>yww_%!nwJi!}3YOJQe5f*Gp)fww=srLw zFh)gS+~hY2rOB@Uh43(Be};Sk0|9*U`R_lsFC~g)e7Zg=|JL5*(0L0awBJqnw<3cx znTR8}wl@-Q?b`67t-Nn6hwEEqJ6-j@pugdmHp-=a%gBSboDIgN+@Z&KuCWzm?2;0m zDj8&i7zQj}EPgumBDrb8z(hB$20#8Q{5ZXO+oL%{@0wuO^(3vgFtv|NH5Z^V*DMe) zni=IXNz0h@C%z%srz^MY`M2lvk9YdWcfL89_rHDxC_lmq=AF#ZJlH!pMz!+36KRbm z>_cw#m)O9~oZNo?Vtw#|!M!i(>rH#?a%M*pzMo4OiG-?H433I*K&0bDyU+>c3=SO~ z98K;Zt|08XP6;%s2LLggr~SgJ5yOib`Qc+OJZ;n8{iEAOXqF^+N`tSUDY2d;xe!=sqmK)Z&KJNSqqiN}S zPx=fm-L@2pUCtVS@UeY8bf{*3GaQg&!@z@BLpw#n(<`=V79vp()UqnQ?;&oyPh`tD z5*FWc2LfVY(y@;G$V=(bEk8h+npwzWfK}|VNj%y1o@Mj607jE~?(+8pd~%K7S~KPv z3px5z?THQpmcf4U1K2OSGWO-;vi02H(p`h>w1r`VX{+xdhwol} zT>@c|+&yp6@dFIp{z95-`u&p`Dx{@H5l3CF_yqij5Pv^irI50yZz14(LO!YM?9tbs z2fwdoa2>Hr0ZcTa{w0bJ32r&1)&Q8263-jA-qbMHp33hWzvTPop)dYPH2bEh@^Al@ z@~)ahK5VDDA*=TvX;AJv-Rt zH|QphFU5hu-UNLv<)4fO#F6W*xC9&)=%g!1>1dNtJIdvI?3!B&_{zmK^8vjtlO?0yK zI75DX8&J(mvY3p$gCzINdn!0<`(hK20jC1p20DK3E}T9%DbI1- z?Q)~kYczw&i`^2bvv>2JRd0XGm@_YlUXV%WPT4Lt;vOo=&oGiqM^F2YXX1>FAY!l? zN6?(v)^KBT`cSpgJ%w=r-Tv`& zkBWWEXEDx~-o0<7k?t zMvB7K^5VIUm0~tlfCAG$pul)DWr@v;a-*DfG<*1rp)#?zJlZK@+3tjv8qnzFxW9lLwe zK7>i=m4j3#^NqqM?K0R-2b|9m<$U9>F`X)u+4xF(f&K&S312S}lIb@)=8V2iYcPeU zyB0Fg4hZ9wRxV~89hVJi9#qN&Z09Mpc7wJnfuvqseE*}~KWPSO$r4uN5EBwSHvJn1DF ziFKH~=%{acpYZ%Y9iFr-72KD>wlFYrti1@zh;eZrN-*Lc?p@8C4a0hckV8+*-T9A( zp#o?asLkLafL<`Y1ayj_H~&HoG=JiXe=;!G2pV(VUr2n-juvZ@ z1cV{frbCP&ls|6VKyXt+x~^0)%3qnwHAm~82+@}g&@6%+15^tLG7Va@HHyc^HNrW< zw23*fI=(o+&kc{cj?n5`4CWu)O(y4aCl3V#j+TWmOY4O%@*C_-X>Y26hzmS&;CvoVF|1Ukmm;ZZzi&|Z0X#CfB z+9j?tHF54Qq~k9nF+!43gTMJA?*egQrX}S_-}Nl>BS_z`9P*wQ%;Zl7`*d1nOgpZ^W7&9@DkU6OF9hrW1=N3bw!2MGC)oHgmM_Y5&{R)m+-P%vw3{ z$6`DIJenp1EccVhqyY;-$?x1vj}&=tm0OhkEZ3JwZ`mR()$T9ft;7 z(GISDjxmCs(4ngSAN_5F_VV= z)iQ>sc*TTdB-5KEqfy2iBpU{oYq$Z5)qul8Xhg!g#^M8vuUce^l*J57BBl4Q7$I~c zKH`YhO`qHKDT=)(bNX!9d^9lrmzt*9yI-K`bqDx{eU>iMbj=FRkCFO8jd#Fg`PTz< zkIiObNRT)6JWVloE+lF+@vggCitT2`uC@n~f48pIfi0SYna{jtYcfCy>N9N92Ljp< z<06LIji>Wg-h_Yq?wA)b8{+7Is!Q$7s}z?Oe?eS)VPA5+nk{3tQ!+(s1S?cqA+zck ziNc+WvC5wwLDx?8PEBvVB4&Rlo8@{NXqDCLXSihgI|uHHa@Bh>8LaR*}*@6yqe-uKGtSjL@BgRy-fU4=3D8#C&K^r z%DK$0ZyXMDq>OTAz5t94?IL3;N4FW(@Ta$>`dcUxE#edCX*S5DKWisiU}O( zMXZ41FXFe2q$sZpI~JHF!uhHPW>co%-Xi7~%w1OUgZsm{ArVv4M+p9hDC(UW#=1$3 zbd+)8YF5I|_KU{LoJnaf9bD4W^Y4z_?M_FJ=+Z=994tqS6>!yi&lWxI`fQ#qs=&Y6 z-pV`F(xJkChEMZ!qzw#$`0VwlT?WI|cgKb;rVX_3byuRkSH-UCs*2rSewUEv7Z6zR zPG{C}Hx`_;?Dlt$$%CI|9(g9K-Th$aZhsy2qQi+5vK-mL2R{v!V!=Abl)qGht)$ME zkg;*(R0_0$B4Q(Og>u2_P;2;k<&Blc+7En(9PSOwB-I`Js(ER7IVpznj5yEW2n?*i z?47QkjL^eK@4`a323a#J4z;ce9OoWpHnG*pV&=~0E#*`GRm0$d1_iQhCdo#7C8*KP zl$*$v2n#6Y=x<#;Cgx|acS0QUn+cb*!|rNxgHR>G_~O;n5>1{nFHJ~1dD6V>viZN^ zn3;SGP`!Sh-U2g2=26nlmOAM&GpNp7kl}$j5NK-R2x)%Vp=2ERp)Wp2zdtKbywj9ZRH;&8Zs9bPtv<%+Bl8C?TBJmsRO*W9IFEg_8yutsg5wVk90y%{!6# z%kb&INLIaSWrE7?6$-4U`B3-kbA4?tIr6D84$3t3gu^n~Z!L z-83W4)|5YpLlcl5+ce+F?UsCFOB^(m& z{ZMf7_rKbd)@K3jGXH6r!OkYOGS_;AVSrmM)y)ybHeq`@LzIVP+Zq#58jo+A)W>kP z7^nqB>PbZc6RN`g_v#N(H&bI4)rg95VP-x-<(NX7ws9nGqAc;Z%!7DKL2C2d$EX-# z6%MggcL^HYpCXsjZ!E*sUm&RjSEOr8K7h+4O_yUk;njYqh6we`4Jp1~RE58s>M6SC z^R`Fbs7HzUw+wVbn9gj=wvw|IMP+ivtStL?!C7a+Ovc;vQC%DIB?^o0O`ezSjtUty z5iL?|eRnH#nM(7M+bI5<_TUchVdHQOs5nD;SDG`iCc!NB;q>s0A~&wqSI*ZS`d<)I zIWw~V)P0RBs#SK0=)yJlhbB7_HqG~L8hJD3d2HJ3(H55#87)X)zQ9SxH8@T1_0R1I z@)*UQn6?PNh^f-O;fi_FAQu$tRvVKee#%2?%Xvkr@Gy)dH6L!C2@zGTrh`P3@Ca z;jVN_iWwfgfa;uYt=xH7z~0{eY+KDw`_yp@_j@Xzc6`M>Jx?GrEEHRc;Aeve z?1;~xl;T5WmH>7(HnwWkK=efM&VeH_ za3ya(`AlCfxj0zM#X1lhBtNw2x*DV`(EAj8IPy;sd}EI+NG|{*0?(yL2WigKo_mYR zc^6s>FXbNeJgW0*VbAk@>fUgxw&r8Rp8GNrgnI1mIywtYG;7U!^%vrvaQ$7@qCTKj zaR{^^O_PP&VlN3smOab+cNv2Ur>>1hqWEyj#fZ0|lz!4n^~8HvL7MLJUX@d~A2~v{ zg(@#n$8?Ikts9zX|7iJ=u-@t;2&mRO4S}&o1TQA0-%WM09eFe2n6J;-Od}5bZ13bf zlH2=1WX3>;a9XG(|GgEKgijDeK7#1TRoR+4FK^$J>y@n>u6WKo{b^kvxBgItftr#hquiHfhr3ZcZH@$G zuV&)ph47jSEy0aVf{S0GC<2yAYcoZ3*+=PchMr&8 z>Ap)2vIvhits&59W>Nuw^y#i+?MK2?NTy{X$@oICn{!RY$9L2gM{S!5yx4V*&#c_L z4jI&wgcfA|(07!cqMeQy4rz8kHj(oIWvX$vFJDE==#7M4Y==`3+R0 zbp*l&*^*)%6s$a~QXn-ktPc!0vQHp)v0d$jZutRh>qe5&Ex^z*HTJi-HdX+GGaL6` znY*&M%~<4c#0-5y2471v7MD!P`i5#?>rNIVzo9>GyGUH>Dv)oY>LwI9Z3XK)TD5u< zjNg3*y*bDdHW1xDk;c(4FHCzHF&z>zjc#=86W{j82$5F(Lg10rxB=KQAEB9-_MW0-B9VE?nUh&W0@fCfParR+`owU_( z+O}v~S+orDYFd#rq=RkjA02&XMxz~03${s|Mge0x061>yE611A6X^r3;eX<dP?E$>*{`LWU-mE*IG?_a@BI0$%YubymA}M$$Zg*Du+tQlQpDb9XU(kIXFs@?W zXd6RaI|Q)TH5bxUa&}!Vj)tuJnOU4_^@yP3M9cQ7@x0aO3XMVppHHwx z_vK0#J|J~>ie}d>bSXSOH)%4m880T{jERS?%EvOV7F z$jKAJVl}z%o@`#_6?2D0Yhf;5IayL!Q49>c)9>Cd=*-NL`vMQ|g z`Ntdt|2l9hUU|;>W%r9|qLaep){X2Fd|U4BnoMs>`HaQcV%(y+qsgVdnoI34)|tJu$al=qGmlayyv_Z%0R@qZh_p(2)D+%780#ChVyo2V(_ce zK?a}7?Nfs^t_z$x|IO~seHL4$lp%=Z^+>}kdTAW(Msu&$U&w`O&@}!V-u^i0_~-G; z4u>;_+EE}+RM;Hx^(UMKk;J!`{r(J z{FS@8ezj%hc2kpE>8M9e^ChMmLv^eFTo`{9Zl0iCTZ$M!@T)ftKbNP5bVtZOQVzfH z<@lu*E-sD4n^NzVq{;h-IG2~R^gQT5TNpPR#(FA@gJ1_Hl~`H?E;lzN5>APoIe+{|6o;WC#c%EOnBQ6=pgc~Vvkua7 zZl@F{-%zWFif%o=_0hUx?WeTW}Jse z*#H&49=1~$xB4aQA`mA&-(lq4{Yj6ki3ZL9==!zzzYs60-PA`Bv(}f~#sn6sv5`1eI+&bzEixP%Z8C3uf0aU>~>wHuO zfFGSd!P=U?!1PiO90Ko%9^5Z0Uo7$3E9|Z{;(vR0O>0FFJfQ&$*FrYE@e6MA&J*pT z-dn-9dP}J1{)7aqF^7$PwL$t$Cl49v$uq;yo>wjd1CzH-GpDo3Uj}M>g>Lcl3$a|X zhp^cH+erY8Jh~+W;mun83lU*(&{QcK_(IB=b`?wnoNrzE==oYpS8;_i-Tf-(GGzOV zk9Hq5Kf#>%J5dR?3Xl5>c^k3`SUA0N*pxON;_jDj*YVOlRX=1)Tgxt5jQ3TbLHf3o zebcw|oXbhm)4%MDq`gSk9zz~tPS;~98wlg|{hKyU zevb?H_JVkvQQ7SpiWeWMPzit{A@}ngcflO1PI2p8dErAtwV!4Z3foZ0;N3{rQaK*@ z1pVnk@nh6`m$+!44~-M)>hGG%&hs(M@q7FD8cRvvOQjz9_%1ckoBO86^q4*fG0^l| z#L?ND&P5r@1L~=;6GT+vR(toGANP3+y+i$^+;__pSsYIfSZ0)7%|Z(B@+yD*0o(sb zF9%vM+ziw7%1-ia13RJsR>bH2z4Ab>il2eE1ifDch2Q$F{D`kt?vD1Q(4T72T%0-1_2ucDjh4n!<;cyD8i*tr|Ha-ZyMN%_Ic` zB=SUPGbX?fMs4XRZt;`k?IzmAF)^2Gr^1Y1-4bcOACVUr>Z{ZE?PCsK-7u4;fL$Lx zztY2;nIG*H#sfr)hTS0l5GWJN)@GfK2iSx#yNWrt_4gj*V(MNoHo4=M&7(Tf1Y`px z-tx#ol)8_|NRM6v#a2DV^>DGj?<<<#R8^Im$z{)!ZH{w+88AKT%6GA=+)2L%Hqd2Q z0h|}GMI^tNxC{(;B29!?Yq1?%-)}Hz*T?p7rmplB0Gn!P~ToZrzms zFmbn4*?W@j^e{Jj3KWq3_2d&FyYLurUDP-6L_u^5$)?Mn+4<&`-bs{~Pt8ffcRT(Q zO36VgWiUS7+kKndt5Edw#eX36=P)I}%1jKlhMN$jD=TMaTnzXZ^5Uy#B=I?yvqP33 zzxV6?gc5K!K|(&6!xWb+E4aLzL5C@3rjL+BX{whXwyUKA9zp z5yNKSw1E}V1h7cqLc|wHe?ZiK_>z3ot;2zAZMtHYGW&;?(O8;MH^l#T{$6IE5fbW_ zth-r8H3I%5M2ZkalR_!Nyu(DUOJA0Vn_)II-KodE+0zaBcD4Ph)Q8Hw!cKH;`PRiZ z7@X;NTo+bgwsC4wBcVon__4k_OVr`Z-;-}78!>0}@Gee5=>Sg#78STSRNY$76AC&6 z?|Y`_eS<@-tE+)PzrPSu0QW8NZqe&)`4Cp4!4ohn*_@6ulxEE2{)rERm0|Aeb9^~*W-^@_ifyX@5I=qfp;q!QUWtw;UeYG^`KvNvUa#b)mQTE8JXQKW|83wM~f z?QrP%;c@5Rj!mh<=}5lHr!#!)JW0`Uca>njqW>^BMySH*#SyG-Hx&df{tAweJ->G1+s` zb52t;YNA0*^QWT&ir}xkH9p;h^&0Xr(lzO)m(ujGKhSG#Ul!#EbM z@@ZGt?)huyTN&n%+R=vPiFIB0707=WENQ?k;2A@j;?;H1aSOT(BeAYsYH7(l*OX$~ zm@zAB;{42HNEjGXCZD^C|HOf4Wey9a3WSByV;c+S4jF^f-dz~AOF7hgzcfVTz^`2O z*IX`g(QWs2Mr-J~A8(jnw?+=F`RLq7&_rO%A-$uCM4HaDp*ZHEEy=q5(LvdV)%))7 zkU&AV#}4sNuRDYnzPswe4U04%812Ho72T|%n*KmFrbg222-rA@ow#4&oZdC+>N-{& zZ;H$xs!PK@#fuN0j5Se6MRFgJfgQZ4w(2Y3A@=tmxNo^}kqVyi*1WhyY!G$GVV6E~ z}=P zIR+*!bGTW-$EY_mPHQt!XFVnsfa!PBCv)Ads{> z9Rfy-Jl*?_q+}8sS%?0*Vdsz!S(RAc6H;LE>o26eZbD7Xx76W|Jt0?b$h`FK(eM;F zD3tUgx(y8}M@XXNu<^#4$aYafUg83nO1{ax>|!9Ao~5~+sp6?3?qund%?yj?Os)0V z3t@1C5Gx7rF2rmK4;R5L85P%PZRYr4QojB@S`B~onQSt5dSUMI!v(v5q@ z(8T<(2Ove5&~qaUDQo277E*sy8}tn(=gj13I2(pe1R#?kJV6HWrAZLg3YVXs zu?31v9p_g1nRAWtqYs!R9{%y?KR&=+7Ci#=Pa>z!<70h`#1ZwnTT&p7QLriP9QDNJ z4*f#TjgYw~4)-IGxp4B%3nZTT(nd-y(LUQ%nkQT zA0^gwR|WM*7`7^Wn0mgLb4}-$I$v{&9G7$D!>ga)9-Z-n_OT;@<1^WgUS}xIm_L9& zzN>7*U7doC^~keQtGdI)zi`At>hR9cPkR|#-dI(2Z@kEdBih&ex!&N+xj zq_y;&s!hEGw5Oo-CMts7yelCBiHQLQB){RD^S3t2ugo1Thh4Ezo_|)F7VpkvU1a(0 zda(tP$?`3tS>XwH$&-M9K#*?Onyi$`C2`A%?~ z_Bi#Vzb=x3!!56@Kh$o?v5 znclt72mB!{-$C*n z)>#PTdf%Jhgas1;ZEaDaC7a8xY)ik|uN2uMVjRQaT1FeYRK4^hTL#g+6o+V7q6tP% z6WHQv5ac`WKKJOp)tz^re2KHVE_U6kG2?W`8!v9#S4du{9@}pmgVUc_)4wE&@~GT_ z({w+VlUS=fUR_=~q;IlhkaimBxVwVTPSS-n8Y5wmHf;Od#F|a8u;5mK;&?*?#fISQ z(Atuyq535&$lqVuDGW&FuIr5jtMovWUePZW!#P3E;i61wXnz z{#`lGSgup|(`%+I?;D4l!aC~=*K3AHDM!^eM@WDA5e>9OJ)&-8W2)fI+@@z9l8z_s zpG*(cIh)F-dx-q}eq8cK05onvkxe=s1dn9;^LE`ZZOSZoQY2|6#PgMSYefeCzV^)O zl&?ZvsGnDB(=!L9@0YxwvUfffz0PO4V`_x51bEqUKCafUyu#&YP9fK3+r~3DPkvXo zZn01f&v~H`E(x{!M=R_lp^+@%9?Ljp%0>KitoA4~M$szHbgTiB&1GdS%HQ{Rv;~=8 zuO;xYr(VE9vHIVc5~Z|I4Z?&eJ3c%$y@hNul=ykIftSxXU3#7^H0j3=4JFzjl_3vO ztRHv?L5bi>Mc<6*LH#;+^c>+s5sTpgy?I9E_9p;?bYcJZPXf(5_TLojY#;nT zO@0>o9%%A8Kpj*MsP*uJ!^g1y6AgHc!p{PZ1{+;a4E3J|-Tptgf#UzeSAA|5XY>}< z?R3(E4XlZO@Qa1zx5nz)953uQXPzF+*H*u~`!p|e=<2UuVh6d2kR&Fy@zVc=iTuAV zge#RFZ(Abh_KrS6uxoGl-XO#!fo&Y<4A65(wry+TO9B@k>-m#`1ut4qpfhi% z=|ChF$N;N^i+Hfk%b#yU`r;>5mF*0RB)esOW;UdnY2aAU7@!IG0bj6rD!Fzd0!2GL z7Itc7G&2>^+)n(2&1bH*FLGyH`Z+aSwZ(01s$iU74n6x~phQYpe4Xg-jPGHDUu!OK& zshWhEZ5_q>mhsq)-c$3>Jelm@M+#|b39~@$yD|cEsXicKo=eh*HI$k>{6CHa$Nj#`P`(B7WjK?yXiKSXQQq;>TUZ&5E7_gvz^k>ZP0Q z2)?EBfIga??&30cZFJk?nZF$Gz#Xyp>2z?B=>h+#0(>Y@q%;Fm20KzhNor>@8_7@w)KL_?* z^*0H1DKZ`pDzd#Y9D_YQnYUoFmv|ux3BS=41>G$)X>G-;w@?&wh)|QLKsx^1_l1-x z>~6yMZf(b7-4ETlcI!TdFu{+$>pyWINCwlD=&L_~qyW(yM`^savfg$iJk*Eh!$Mo` zouyvcX4~{OcN@PO`w(v4&Ga-g88Wt3mnJxDNly)Pt)VQUC9DVwHkWH9?p}3z%w7~Z zRBz51I`K?ASG(KKwunC_&4_JXAp#U~7jZX1y$vzs!+7rCkH68@r(+$dI=yppAG*4Q zRMlqwM86iR_-gct6$;5ch~6Vl(KGLEtlu5RJ(iYsek!#REJ^~K#k>za$=Vo;8MW)h zglsJkOThLCRU()1G}u?9IoxT8XnG%GJE61At##R&TrqSs%PVOJc_xc}ImvyM2^7e2 zWDhzH=i}-_jsc70cK9g=YcAu#BM;6`_-$_En>8$0es?(!K2I-UwZCKgNbD-dN(|kJ zd)aRT{i$ZNlgbw{oa@zJ$m)}8wxwJ`nZEeBQA~$RRnew7agOlw=U{A)rBwCH{WFKZ z+diy|X5|>5H$3-ql;X6mh9eFg<->^tSXvTjs)s z)a2hgk8sd7>MRS~B< ztF6%$_F~7>j;|C%;|lOpo`^vk6bfLNkH{IvD$C6%$^Eu6fdzs6U{b5TRYmoQ7;ERF zZ~=4w%uYi&X~a~7%n#qn4N->x(<*IJr|)i206}k0OT!+Pe{~t^c8JOz@fF?h7n1xJ za?6H=a;U#uuZ6@LOW;sb_4D3D@iVVFJ0Id@MhA|Nb<2g8rn&zj&g`@F?Y; z#aeKRj>N+=YO4@1LCjzYH+WMg6TEtiIh+TA0<0~}IzZ|pKz6iX=+Mqmc&IuyAlJwK zS+V?56KrLz=eqBMOzS;4@ac0Li~VS}kOmthA7U=Q&j6z$A$@Z9TT;a1AXJa@hx1C? z%c?%1lIp%E{UQBw@tt$sK%HPMcOg?L#%uF83H1%%52NJgc9?yJ(B0-r=deWkDC^c zcMFZXnOA__i~F}jmj%aQp8>-HY`-$wk|MT})8A1*4cNPf@j5%My>UW7-y9ED$+6o{ z*{!b>2#SU96+xpiEY?xWdAKI^EjE%?FOL5gJfT1-vxl6k8|(MG^2`GPzv%7pn%20mmGfb*He6YcxS+OZqEAWo*~V+ zVil8G&gXAfm5w2sfEefZ15wXgOBlIvk@S$E(n>PTIM6uasZGH*JkR>RA2y%7E0TOX zzrvP%&Vm^d!L^9(G(q`t4Y%yTd!is)>|`$mkz!-Qz4V{4A%r9` zG#Eo}Q1EcG{y_W8C5hU(d}U_$q$+&c;_3t_U*lTzLTlo;Sl?y=c4 zo)>oW_R3m&My}zkLMGYYv8;xA%G5&0@0DuUDKXS@5g=)B2JcWzKx(F%jpCh*<#`akyaOAL;0rtU`DGP;L4a}axy)FOV>#+M$7#m^ zz?$1Vb>DF}0Ywv1BoFSLk5KxlC}nOHCYZAHejPm}d9c%on;XU3JPt!FkBZD~9Edb~ z8)k+Fm5xf*In!SYu&*9-oSTI{iCIUd{d-7H9(EljB{3_OliB>w(Qya1%!sDct>;u@ ziW&JHy&%jQ_%H$)s1Jf)+dJ6Wjr>V${IX}XvddBTW9Ea;pFL3;gfT-Ci90i9WFQ3y z^QT^*D0VFBk{TGYV>A;LvZKRvZP!JwjNQ1?iAmK}VlpukV4_U?j>y}6fYy0%?tB8fwF?}*6H&h$vT)}t8tHBg%tl8@7KX-BB1 zZyIV4v0H3liRhPK>}#zJ<9DeT-_lT~P*1ZPLHe3_WFo`D!K$T2kVx-k zQcr_A{m-K|Y{(1mlOniR-r4t>yoCaK{ci-2NkqG+wNBj{$ZrWU5F>;ejh7B8)0@c` z865khR60ndMj&ru1rLuANG9~Ym*hL^R+#Gr+h?>pESqgTxRWe-A}jutVPtF+~|koj?4A zd}xKFZ=2sKXE>d>P!j&;kdOKk-6MVM+j=#h5x^*EE021b0=qzQECt$rEzETuY?s58 z-#aq|qX}1ZeI?U)0(**v*i4C-%KT%g)`*bLxNA1Ehd1o&>hj11q{*+`@W5<0e7hS*bXMo!=!kMMue*lNIWD zU#Xfd` z8qZy$H{ouAfeiC(4q=7z=r}1CQ^??-no52jUcOmy;iF4ZXv94WBdXcpr-i?eQxRAU zuntk6UcXFfwfIm-P*^OHbhOQG1X`JBk3}uXMEMK$@1@?nf$E#<$&*Vu(Qj?KPzXkk z2o6dUC5ivu(z-Co+u~2KQX*utO+t4QoM36!Y|FUa?kyV+N3$bNQ7!^ksB0y_abU{e z49U~l>U_hyMCZpo;bK>YRm_duEpm6SCe3tQPfl861IBq43+$Nw!n0Uq6T2P5+MaI<J$X^e8lK|wPpe?fo#Z4q*r}wFDcvjhJW|=$)`uh%Fd z`~NEqPhDKu=^laiEE>L_)=u_vo%prUs$5ZYY}h~ir1Jppz+@#$SS1tZTekFcw_<82 zW4^FqClb^-Xo)ZsQ9VV?09A~MaDL$9?4ws~+_*(}TA$;f@~!j)@)zU=#L#g^WTN`%RCStiF1G6_KgUA=Zdx`GsC5S@JFYGj>-t4^V9*b%TBL% zeXQGi${b;oJ%p0wL#&PCLP075*gGl)>opp?BuGOLuo7Q=0%)ca&eoIaCC#blt*n?6 z?H454G9y<;?s<96N3wp7wpsmE_!hUBp;nrV^%;fVHtj>zr=_vk6*gbZ=Vf2Pod>{+ z3OzS0uqNz0g@s5Zh$db+p_7a~Ti4c#88b~bsAbxGt!LYx&&wJBk#uaC{*){7frutM zQPh`lyg+`FeINa(VOZH!3oPx>e|K@~x{~qf>SHOFO?T3KVd;I!{vFtcoDNi?!~w+_ zzx1zjzO{AeV^elW;INjNw63txozpRyz#i=f59C47x(Tf8hEMMI$N1;cdp**!AL6Vh zv?;j{^70Zyl&&SPUS62lvWI0@2tn)7v1p=w&0PNqt`m{4$WMfW*fNN(P>RvUz_3wa4^s&RuNj z*2F*IV;6_)voScTs2&lYmxQq1;VArQkH6>$!5Z%P_JjBJ)ARO1jEexjD z!S)U+*@9l!E=lvlA}Cxd*@Yxs*YVqjceUv!m9~=XT;PW-zS#w8nHTXDakF?b7euaN zmKf|MrI@(hgT7Y$PU4>T{d>j1sy)Y>C5&3Lg|i-SdS`4yfMFBA#OjTtm^64Db~(3D zvR#&TpPm<&$TcmDSybJ z6jdUXEKFh3eH6SSs=V$XDtPNkL2bp8x|mPzc2|aqJ^2p{3p;+i?re}@Boz@yb&D8| zkpJQvcIri#?U>(h<#iEunG~|-?$D~-i=^G&cuqEC+CPgA#qaQ)3p=M$g;%ir3`3eA z1%>r_NwXz2W#RK$XFk3?^B`DK^ZgOX+ob37K^@4eyZRCzh!ZP?ol8m94-j3WV%`Oc z;o7qAoVgx8D~6|)O`SK=lM(Eci5ibQ<%!b(*xV}|f^j4aN>RjbS>X+9>9Mai=Eg6N zgh<}nE9GF$p(#c{g$-nMUT6RbHBTBEVA#w_aB#1^&Q1Mp<>IQe zdRK(;h~B*$ZFJhSl<$pYmdqmvSq}F`*npD(Kn^eU87%oqBV}c|%F&YkvLRJ;$@_i2 zXw!T9*f$Q%qB$bTYrPWYe=c&1ny9?cynII}+8EJ-e(STb;G%Z z)ztgZ9IhSiVAJXOX$Get)ob9>8sX(jE2*B_O?~%?25+ES6LYxNHWZ5pFueZPFkfC? zV#PF#_DHDJdj*yPB8I5Ou3EM1Tiq4%MI@i}S3RUF$92A&HGO*2L#}^fGFKJ(%PNbs8Kx9ZuoDXshaL9gWHm zWmf71(o_Hld__01C4u;lVy5W!EuMHc-C)D#GUy%b9V&D_R^s?o$Ll4^(M7M=YW}H2 zCTr%v(^|j0lZ<9-7o#bU(_BM~T*B0vee@5l`>uQtod1}P1om^rd$qCRmtE~c51Gr< zq%lUYM2jl@W%WL57~xYrNzR5J7w5XxTpwd8BDMxw|}5GVeU1D-hL!AAAIIu z--?e zzmpz4%}@vvq35;9;zxNX8@=zbkM2^fobFQLTgnxLY^OH_r#@)s_1B_-Fi63y5Oll2 zF)>V+Ud&*nxlSpLfho?C1?g~!x9D@#4_V_@M@`bWHre!zWdvBEGWQ)D|H~KUe|n>Q z{%<Z)&GmAg!eam|ZMEVe>dc{SFJ_y^Nad?-z2iXw^0Lb~-8 zvMA0uA%;~lH9sZjlMi(+;-TbnlH=YXk~)L)5Y^BJG@r(=2C3>8XIHJI^GkCQX-!n$ znXhKgkBQuYBA;>xQ)IxOSlcO~p#}F!?)Jz0o-Naj#c0N$9;P(X`s}#N6eeB=_JV_>$!wDk%@h z@N6l|CK+p|CrK$SRF$WF`qOLhtyEAqoxsk<5az=UkI}f>bJ$fH)GJ7=VY{b%!C{pqbm%uTe8ZL1gn$B1Pgp;;gx1 z)k0_!16n#L)+#&BU;8*qi`hO#7IqH!$mgP5iD>KPR6;jn$H-G`5Wg7 zOj2u0#>rs3_qCAGW`}r`8&I9`AAXo+V(1TKt`^?$Y4aBNsc(avo{c_78O|OpeIM9= zA;_46_1ooGMi~A_NzX=Ql4DO85;75 z>03z;Tx*#a#IySZUvcyy(YY|(oETr^X@b7O;>icj@ScJ!ZR}i=XfZR%52M$01OiqY zxhzErIT~jACt5Zp#MS!RE7mg^ak*J^5sBCfq=-|dxt>e#6~lR;S3F{e51$y%d(O!& zwB8FOOmsXDI{fiTe?qR;XMEVQ=^RgS)aSo267^{S(UW9(%C`9Pi2YP^?v$&-w;=c1 z5}LR6`q&%ucDX(7w@t&f(l(MNKEv0{qwlup%bH!}N-Bj} z0Op}0=p#xy+a>tV_P6pKF#hqihAx$QB%BlV6;J1X@UU6ir?|~1UMToOZVS!NnMXXW ze>6!V$iAyCfI`wz%G70yd4$`Y`c?6p&-P;b8!XF;T+Zb}=W{95Y23A)9RgAeQhWb_ z0JC;Km7g1D4$xr<`g|rKl@3l8ye&&l3RA)*jqW&`&(h!VqLHRib=nSbhthmIA}VSH z5+RwMF8@%w}Mo8_NUzMGI9)T^hK7 z0SB+@s!{2-kL|Pml0}02<-^|v+|32M6vkHe(%KA_bdq$+CDW+KKDqfh#rkMn-mD3t*1AD-D^F*j*o)JHhE=d!~=oAbG<^#nDtp3#TE-KGQfAY&HI}Dvc+(oz``;79*u! zE~0}c+7gsIMx^LbG{CxcV&hIstv*RY<*j!;FrUbr@=>U z&;OW1x0S|Zqt&ulfmJ6nRo$s`wrys!Pn@{xnVzE*N*uU05md?T2wr#`=4hyam*~Bz zS31xcn9%0!*frMz6E}=FGazxZ!h~LCs^eD2<9LO*KAJtC`Tw09My=0{_TQ$s{a}aT zgi;_V@Hht811TIj)hxSyeHmJtwZYT-6i_41q%f^RVQt9%&!X0+S7k2Y(!gCdG`-w`8mWLT8Dd9g)|a~XYrv-7){bvH>A{=UjD(cr7dGxd|$aKX9_>Qe|& z47UVc@YCE)Fokelh5i1XY#jE<x{eNL7yxQHkukDUkzaaky-3E?Xx$7_lu6 zU(WU52p*33NvPA~|Ci$H`l6>Y%3vPT^ea(C@SPJ^%aHT;ghlXsqio;P>dp6sQ^*YD zZq$Q~I0|*a=8x1ZmAIm?!6PhL6vpZB@{vKcr+-x5ek$Rkro4&nRD-Cb>c_BKyY+kTLkYLC{C$C;hMP6f7N5iZLH;a4x|dbDesL3oon@sgKur6DFekTL`XwEF|!|VM=91lfvw? z@qT0IHLlzUXGBanKz|ad=}fH|@QJC1z-J|dALdg9*|n|z+wQaEbkNJ0q0CXBc$WlS zW6Z#2XC}KZ_jhwb0nE?BAG>3|XLBNzRI_(^Ijvf#ows+C;TRPybI5Q!?f$c2K`eu7 z6o$!%HSb-=T5}%kW(dU)(hbrcLo%#4Gvo=KVD&STO9ER-SVJ%>%-~e9q+zkQ%7WCt zd(6j9Y?7K3-|f)NjY}baA+ps)o+H;Fy^xIG!+L1vu&2`Rdha}9neb75fUTI@_X`mY z#vV@wcbSb#+{X5ytD>n+OAwQ4r9@wscqV}pUqRWw-+QQ%)GZ zyYaqZsF4ANOTACr7I#~`7CiT}Ve5@)yQ#bs;Q(2!2b<<4}EGyblMMNNDA806i&FcY_*PbAvv-h;H3zNHI{SaJfz1>bwk|H9zByM+E4T}xFAHti?&n@cwecOww3uvywG#st^sr)U6F zOQg9Ar**JWUB;U%a_q7q-}x}A?PtkqI<@z%M|QsNjTYJw`%*(#CrWLVqGbmtG`Qb~ zAfW|wD(K9l*$9CPQG)kKSG0}Kr4v?+BCC#ZUjo>fXhGNf_v)htAjQGF;Sz+oi6y^F za!NYC1rhvK?XW|kDGjWQbi_}9vG*`s=hp93eIa7V=A=ViL42!Fy5id1`y8~nQ$Blg zcBWZccTRZQc9Rui4Rkpk;ut`hIv~b}^gt%N2!nS3TZ5~R=>c9yLwc(9Yqy-R!k&3G zsqMsQ{H#W_z}oPGp)w;6l$GcOthIC+M|WND;OQd>EwBPEI8n*q%a zJaI&yW3J1x>aL^7+b=`4%f^ohTgTD}+E|0M+tDl6m}qjNeQ5sVg2emU@GJtYf0inY zs_K5{2Mou1t^F^>9J8o8NeyRf90`N$+lc26v;7Z*^^~ie=x{(#<2nA;rn+JQy&p#F zm2Rj|5_C&r6(8!S>?bG6EQ?d?V1Z>vszW64{`S8&VcreRqP zfN_`@px3xaxs${vN12!!d^;PpXnvQ=uDseuqw=Ttv^^GTQzX1jwwNJfvTz91_d|`J z|MCep;?X8-B?3b?lJW~2r$hdpVWIkY|4k^q+vf15e^LH5n4OnxzM5hEve>j5=_KvOuq@pkDCGqnxta^IBfko+$=?BWh z`-HhqYO(ytoSIG_sPYpJ%tIfV&o%f=3X4-cf`%ygEHZAexHya?Ocea#AvE7EK@Q9vCmUyeB*!%U* zjg>gU7os+18ck!56HDrWd==xm0An5U_tmedQ$iJ6ju2vm4PO;@RoI&{xrPN`SUqS! zQ6*NqT2!9+p0SU_m%TzMa$5SSUA(lte`rB^a4Lo8=+8_Qm6*}zu|L&nC5Cu~e*dn- zK-8QkYVx|*%zdHQ{XAl!UOVaVFwx&d>@qNg{Hrmgje-* zB%zT*AWH=DmDiPy7M^}v={oBaEot#CqerYaQgWDXnvZ@@TdD-b6NE(MWAL$1PN%ZO zzOADK2Zn*9Uw$sk2byQy?>ZeQ1wxzcJ5s#3D!A!f)EvDhEu6Tiw8R1^gRr&>RoEEJ zOTR&m&AH8PYlAMnc230&-T)5i5Otn!|M#(19}pEc1%ls=NP zY5>kq92^d#X)_~6*#7Kr4&DYC&jxFcys@7zUZKCpJ4L+7D92>}n7K~#_Wzds&Hirz3G${!fwh`8 z!gdpN%BQfcIe3()etH470p=#&5+6?3-o3W<)ptq7=fvWHtNHISE?fCB(?1qyYp4gc zK3FxePyRWOiG()%gB2k}mUNNz_y5|oORv?f)xUmg9l_d|YT-%aIq@{-pf=dl~XW7-JrJ@B3AoL{M(IRRTn`5bZ5dIIzq_+isVEm8YM8nvya- zFzjQk;2PG0z^)@PpT+Y{`*7S;dn2&8K<{oXyZjZ79plNB@ zc?wyE-?I#W&>%J(F|g6%-B0%0N~pub2Jmoa7$@=Ci9wTN&2vv*BQfVfM zvJ87d#vz#-X*(QGHW4y@njonS*+Oyp=7eh&dF4fgvTWKZW*HpIp?jI4j^w$}iT{uCCl2P+Hvw zoWr^u!si0_3Dt}XbT1SKJ(Hn1RKz=^3&1uR=J%2V1cCve@aycIe)SbC+EI028&<&o zI3s(&UFQ=7vbDgkQ8a9d0)QV$mb&o#sgH?FNXfuC0{ zs})|_e7;hip`O|T@}`Ws7~~b)j}a|G!>$5_nT2#YGHe#<*N`MCGSegOSEQd;?qule z>sxhjX1R~Yh;gSkgWd$%k~Wm2Xt9HXK**QC%ezIHa5;)kmE;7L1DYn`ZDPFt)~RjF*wX;ge#nc|LN<=24N5Jw`cER$0+G2=DeT5ZS>&w&ann{bSmT8+4B)))WY?06|?q&ta? z(pTClJ8*=K>qBztOThKIF1><3_N}iE2neVv74OAY_+*Vq@q1;tazX^5__z+zW6+lr zR)l#Ym#c2f>IPQ(+V-g8P7a=Q5zf4j`|)EdVFoZ;0(aXq6cc?UEVQM>xiBc4bs$=7 zJHL)}7tS*9@ptC0DfcUjyTF9Bb^Zd=k!!wGAT(NFKkT~kaD{sGfB-GEA!L?Z3B$bH zXx&JH`G8?cXy)t#N_=t>JumAy-F)4H#-Yr;^)%E}QnZw2WS=t3 z+a;y6dAgs2`x}a{W$1J0T_8(!vN849%h$S>e2+~c={?(I#amW*pVlV@Tr&Y!hQfRV zQNA3>O@{b0E=_H7T_*4BR#wpWZ7((&2YQA;Ift{2zD6thP>Fl|PY6Zf%rjGvoS>h3 z2JdicglGE~LNA~sXCwX;?ZG$m8F$qF%zBw#-%Vm22?+WJHu{V1R=`bAm~%;Dh#4_a zpdoY|DP`2v(Y2^@^clGNAw%f-y%MIfj*_TPvhY4r``(+C8F%8Sj`x47nfLOZ{zCS- zg1Jd&ob0d*9TC>sqqd-{T_M7NouV%cQ_7Agtlv*BF^kVG-ASfSfT}t?_=!gdCRvxI?K=BXv=wmtFE z5JGd|*BZa=oXXVDoAQVYJD#u7#)QNROWhUNYvN+fj8ll0MoL>Xv9fjW73T#NeOY+M zr9zbe3%&a3@Znsjz;Dd_^6 zrcKY>*2k_8MbHLKy8O9(bVK(=SeD?#MOBSx2wCD4$9@%s*$j05W|90TA5vyAj3rW0 z*Q)IP$Ef1_i@&d%LAh5UraNTAvJQ!08biL>UBvezATef27G$k32j| zSB1V~Gop@pDnR*OzD0nFM}`5h%{N%s?aYTJ2E~hUpL9Whd5QnN4NHXGj{ru~w}$@? z$CZ8k>zNh1bgidoT=iran!PgWu{vqK``!RT4Wa(HL6rUtB$YLT`s&1&uZg9@=r^6- z_cmFkoY(cP99ZJnb}Jey9-0PNKIclqsTyZI#8;v*?|%Wsb)ea%lThv-G~?oASY;U|%zQF@`DuKL9=0c-+;-j@YA{A3nzV`E@?u zEEE-7+jhN?q26&TftHhz($8Pahxs`KK`>H4Vhm1gE=)s=SPMsdWZ0{r%{ALCnWUm? zd4ZuJLYJn5HMPBcSW~id)!#yMv)~X9N>C|LW(nIy&%YTpBlsUkDHiq(jG6=8_Pv#E zxfZG|SaU&iD9E>g4x9$JKS;49v?rc-R1j*A_&u~gz-h}ksBEC?EudGkGV@wYVhSMz zauPR+`sq!gpC4|*0LPH5a;oiJD`KmGO=GR3cRwV;zy%H(WuHCU zqeVvPMr@%H#NKNT!UW-BH(!`7*7;-I+B@6DEjiE);_V?qu6b@bs;sKx*KWsBLByth zW&Ts|&3){Ox*^swR2s@uP*DRIk6<9s(r}vm&$JW!7_{#sYCjr(DLKQY^EG3`sXa_ zmOav5df|bcb^V=6@w5y>TG6L&s21ylvs1s5D8fWX=LL_GhE^r{@q_S-@oS1RHBJ2( z)*Lv5!{8hZt{k{llGs5AU;rTimW<-jRRWe1KT^xTExmQ{tMPUA4RBMf2(IkP$w9ac zwk<$J9~!^}BMts1Xl4;NoZ~EUGzWyzfDfu`==Ehr+4Ks<_hx%jLh+vBo6oprq77;O zuIq9_cEDK;+dppD>fBPnT9Qn^i|XRpt5gOzxG(?KoXd9vZfa<8+BDvq=k7|G+yIo8 z>sk;hkiUT@A?C#fHA3=VK#X)|C%$$UEv(--Xn&IGpje~P$5YMYlcO7MwLhU~)C^%1 z{A56eG)x5&rNahGg-@->d0PuzQFUu@Fi{cuVnv`n&_ly|=W@1r(rnvZLZJ{{DyJ${ zFE#24D2sr`5*StiuLC*EfmXNv7Y|i+Ez9qtLo4iK57y8+k6c&)VS6y820YI zTiqUjVcr2y$a$h4p{xTJ_OvUo*5~ZS%s0#77eD$RF5XD|kU5#XYA|~DdGJ@D_Yuk<1&C;xRed&c5i#5<10ah6nsK$~0Oo2!) za(*NmZf>Mey1x63Ntn_q*N5COzm<-F!K(wi{|$&&e$Y~zT{wNYftYt`DJhQ%LBfEW zcyfS;uYg(Yx@l)|&f|CHyk@Vi1K1Wl9THf^T3ZVdo3~^SBXGZmoHgfn7ekX@cem#s zY^F}hJ2+4K^FPO)Xe4c4pIlZnwSewL>o1C&gR)!nRyORK{ZoRPL_PxF-!D^?=xBha zbd-NPtC;_IF~11$TJIT`8NqH>+>FMwB0oD-gL*`q+W-PV-sxbdIzQLfurEYlB9|ky zoDtKgmNzwZ#Fg6%Kf1&=`>MZbxO`fAzr6NbIf1wyg-siLDkn{^k(OIJ!B4J4`pm8X zJTWizlWoTyn&UJ7hAY{?+IWWmjbTBQk0VFo$EtodPI#;}n|>wl?mKH9ji)_)d)k~= zM{8?#R0Cg#=k@Al1K`ELjY&al(fO zjGi|I*aid`Xm@DJz@dRD5o<;LH`zqU;u2$4Tt2sAAN~h&tEIE8v&C**=ugq3yUi14 z5)4-C{Ci!4%==)rr)=azm=7ScF%idbRW6_J&FQ*tr*`_)UAk7Zry*gbV+WN#Hg$md zKE2vRO&q&9WC%t43SxN0a{!^}`GfnjOWWGiq|Fp=)-PD$T_Lomk+D!~b>>6BJqzGPXB_h?LyIs zbipwUA?bjUzDH&LRJ+2W9O+}hR~RN;m><@WFFAW&F|Ak86*3s}2{Jay@RQmeN_RWf z92RM)cMlHV?REZfDb+()mA($-122ns47&~7vKyUuXrTRY4Dqa zD7}xR!8Z?XwAvJ4P-lqS_DhfMy7iwrIUH!tIS;*nJNPV{48ZdH3XUXf#PdEeG5ig> z)jKUZPxa>n>_ZHZ3olEEBp2W67vrftb9M2Ti1W!!zsq?tK9VI`?k3q~iZPkb?Bcy3 zoS+$wh&`v(@Tq|)uQ`}09H~Zr_KVGXlhPdg>VW@mWqMkHO`2P=<5bsZ(>Ut9XIr$r z`2!Dc+NG0_O&4@7A${i5vHXA!b|%u?yR9>3-s;`z>SgoL&8;G{^!e>PPkA$1t|180 z@A-MxxoRJ(=NJAIf_8=x`6Wf30*W;By0G$7=? zlFw>*uUBiV7xrfUI>T`}Nf+6t1GeYdneFy}XFe~S>!+2J`Qo9nOJHu`-EhPwbdM1Q`_F6@;)FyEG_|oXzkE0KRn8bri=l_DX0pS?GltMX68585`a>t|_ zSxGrGH^*uzzY@3P?D&|##QMMxz3!C8n>blTXYR@f`TP6f&rckc0BVV+E^V2{iut}( z!lJ;*S39E`Dr?_$@0moxol1EFUCd5|dkXmih@1KU)|AkBdjy;Bx+wllb(wK3Tz|r` z>#YBTo#0V@Dl|vCO)W(rPO!E;egwqJ&gM;Vh5P9NS^Zu=`m4K7%Ko7W%MI>NxrQ3~bamPpi?*$ps zZUtpVhrJziQa{yfO`o+h36UwRI|OkF>T~o^1^ZON{#y zbZNhpiPDardD+5NccQ(t9wwaLH=0zUthAn~msMt~)t*AXXid(#X^yjpd6y~8i8 zx7ObG+$y;g&38V-(JMnXJ%!e;+^+qKoj6d2&8_%%uTas`%}rFnFO@%U ziH6RddDuwn4s9{ShlZeX4J?OJ#NgkO*Zyw|@sJVe`D*UMmMK$#`3`25QNkv#p|;{A z1dC5IoJo)RRyk3_L+js$sVNew{7$l z$7E)T%X_oKQk*wIy_0weuQ?d@kt6mzX|r0tNq6N4)e602W4E1g(O$6NoP>GOa&Gzn z{B(kN4`)VU@>3ZsWgGCrw>vF(ks|4o0&6RqT3;^2*zI=sog4I-gvdVmOYh7`J_|k) z>f7J=qr+^G;Zi@PaXVt7bx8`EGa|IYE9no%z` zjoht_G)CPeYp`u;;lLh9gDned1v` zNf|D5d#=lmRgCSjq=*c~H0MXeasbpDn-&eCf z#q|P+tTjxPsJQRIEiW6JGesyym_p#VlB;<9g+(T6kR-sUY+^n<$tG^UuS!arcaJ(RMQswiX|8}+2lOn zPVg544&>@q{GA_FQobi>J4JQI6CpFI6jkW6+M%GS4*?PvGe3pkyeUj}&9yE#lXR6ABer5#@;9BgiejG!|$`eS+92{du0eM!e^q5b!z0~ zwW(snsTGbNkY_&ZhylFX-%xV1{uxBS?FL;B=eJUxNM!J)7zZ4&blAh+Irp+*|6?QE zjs2=64_47GL&nz#Qfn~EN`=e(Qo0E{J=1)Z{HhC93k=PgYTymSk6Cm-dX>I&9IF~U z;W_sy9_v5&Uh%=8sc??^7l~{7Kx!Eq{Z(i01xXDt1}eu(7t5ymXICb@l4ndV8G7_< zD$*}WSTKkGlW#XrgFg=Z;=F-CBiTfU*Dd2)KXw$|I?K;BsN_YrAr=Zm-taNevS|HP z@*4y@vQGoG839KJ3ZA+xcEPrUrmpzanT}j*eQ~|5y}S5N0dsR*1*W~o?qhi3gxf8a zcAB3;1i4az9+8fu3C|pXF#{*18qA90w}4-D?zfa!@z-9vXJ{*>1?zeTEF|+O&F0e? z^IeC%C`Gqhqvl|~4Jib4UQ1i9W=dYZ^t8u`PL{S^o~5?%IR;&32syqHd|hxDpcBdi-}K=b z+?__V6(-)?zaj2@$uKNu;SSH46AezDhvwAv2q5HaOzan15we}p4c!t61b75Cok54b z(8>vxp%9-RUYU~MPR}N#{ysZvH3u?s%tgz8#!0$`t9eR}@$IX8FIg%B;T&rk;iccU zxRTlL)<6yQU5Mzsez0LWSDr7&Q(loA5R;wPXh`cCH!20trwU|%OvM!cgRXdqCDq|C z87_w24YSKYTlxjI>ss3^q4&9|;;4(W@} zDz!4Ldp0LyD>B&jrMY(Fi{Hxbu3FGQau-pqfSXG5&(pYpmNE%rRnq5!vpd$jIOm;z z;P;}PL-+TY?a1^Nc@^e&z$^)JU?ID~aOK=9j=@agIkW!hFH$bV5&BO2^w#5Y4C?G% zyFul;^P?506Uj8XSc`JYh)jvLAgXhM2Cf|wnzujYhcAlLGxt_YX-P2@{mLXtmXt=RV41e9Zb)R;zQT1iS%}|sv+(S0@mY8w zz&f6B+Z(FDHG`o!{JueyY(FFiN9YYzr2%V#KxoG2{<90G70^7ZDfVlhV{*q1{652c zkCxjBFrqsN3WsfS48wTV;nw}@T%9h5-KLtptxOG7%5#k|QqRV{)XhKDaB;=uq2JLV zs%^K9PWj&V|J}xf#nfZ#cJf(VIJp#4Dk>td6--?j{J|LSYRq+pDOfYCFOw*fc&d^~ zw+dI;R&Lr7<8(MsUG*(OlzU$+^$(6>+IUc>fi}?CqmvVR{>GGw9=l@wy=o{E++1!G zrM87K6F0a{gGEalmL>IuTKK{W-iWDs-ZX7ZxMQ{w*OcVS2!)s+W+{(+FB+vCy;u0Kw!)GdZ{p>i0VYZ5yW=R#CT});0eK*CFgGkv>Em4|^b}Jz zGZTQ0MSsiEloO9O6Qq3%%pEv3kyvlirAT*NefPFYBIROp_vce3v6t^Af~4PHyg{e3 zHzhKA`Mcnq2j;f-ZZNLjoesvRTBhA-9RUy6i~L`!N3|!M7n3D z0QnV1s=nSSARH6wptC-G4L z@>$f^$+j5_!BB0R>&rHDb~Hn`@zX=FWn)4kIls%d`k>@|a?jZoQ#J>(sHOIyN9<*I zs+l^IjkVg1GGoCLi#vCMzQ865u%7qA-yr5td`hhf_DV8=ynY(;SJKkl^9lA+k*c&c z<@7hT6cKltfQ9mx;f@Vpy_gWnJ;r}kfz#$P}w?!Gk8!kkGiiM5k@{M zW9uq&SoAL4j7a7Zs4#c+>B~-Jt{lCQL*)k9r$*I^y!N5#MS5*4uk?bM2hdAbK*m$3 ziw=H_UrG*??dT3VnGV*U`LV)#$Kbz#md6{KgNS|mZ5s${J{_SGDNNlsB;hqNw)Bjy zVeNf1`}@VuD^9IXv(JZ7i%PlWj)>db`EzZS3gY3x{Qxlqik%Iy;6HkKlzeJkb^l{V zvwrs*3t$voJ;igQud0JXCx$WI^f<+SMcOF6^5HU>xn5<9_RJcX3%S_V)G`Nzdu5W; zlan(5U_S=MF#u2c(-}V`EfVvL&?C{)oVq=oRGxAv*>)oCTG18{)hKgrU7TVHq&H|J zsE;XRSs`*#r(1T!l=W>-$f$*;koSs+ID50mtyYUriypw41k^>;t6 z?vyDE9GVa1VuBQDdP?&eUVpsfL6z(BCUy3ufmjKQ5eQ$zmVbl6VQajK;X>%d$`sA; zlZb(agR_I0;xfaAFB-^nqzGXKy{n(BCP zRc&WB0C`?1`p3xid4SVQl8`q<9`v!kP`M0#pv{6z&mhF| zAV{YTKGZ#zXt9=!zUhP#{<+Z9?@Ba%d~$qoZ+)9<_bdfpr!%(~)xT%beY^5_rZ-Iv zk_88D2qfSqgmcF*3|Yx;^^|Uj_&Lv?mykjm!)MH>uYPYdU%#SoaX~3wR!}@EIOT65 z;lkW6HWk86qIK0l{_OHv-{zW@u7;GEq?p__K|5+K5albQz}+$^^FD;wxdpf-LIg9S z^|q2_ul@!6mZM*hZi&^rs@?ip?agRSiF`DZ)N0jkC|jQEooMlxuO)2cTA25S=&8~| z&v|0taK1I>h@;@{l^>q@^sOGfEYMq$S8A6Yyb z{P(Do*1|kbQ`3L<{H71I)OWT$L-qb;cD5_1=8U)oC$RNYkv<*%Eu`d}qnmT0^ybC2 z$^LpZlcNTGT;~*bg39xIUvYey2qemF99o4((UiJiCK^>-cO z`EC&596C71DUWS0vPmrSmo)>t6A;`I!!T>Uhh5$JcaLz5_yf<=3LUmvxcc{s0%+mI8x(%defi z{I%?WFIe$%Bh6&y!ymU!luZr5wYvQu`hpptTB2CC9f@%m@4+?8VZ?>z#^-4C{cH<2 z$rQu)z7pwmb%Q@R^TajiZaM}Sc#<540A5<`%S0bSa$Z-VF48!YWPBy>Rk5R;%i3kA zaRSqMz9XJ!uC#KluT}^SWS`7=tkDT!e5S!)$+JX=P2+Y>IG11!Z-iI6{iwk8yLpCM zW8&=Qyg0EzKQ`klps|yD9_xU1yfoo&ij_(5@Tp8+^rEvDpa$jV3~+1dC70>mBT4?3 zrEc{fNUCngb`q8%JfUtWTIchr;9@)B(B^{UV`I5eXr@G&;`s7PEaPc9j-gM5C>{Dc zm-tob0Ba_)Lp&ZSZ6 zcV+*r0+jS#XZZbA=gA$%)FdVGoP#}+1jgGf89VZlZ^xj;5XesMu>1^$$|*C$wP?T3 zS)#2iy)xgCCGRqr5Ilbs0(oem3%V{~-BXxDNL=5-+I1o~xbiLhp*b?Dz8}-@+G$uB zZOJyDH49R0tuecy6n!^p5Y>u2JlW)U`q6fTEo8Xu23*DIT$|?SWLZu2ojBp3!8)~U z#(R}El*p-90XA0sTSC%0rW8v=@zJ1{DAt(ljMV~nN zwyRZ$iqFet7x7a`b`+iVNpWl08_T6bVuyD0qy>`ikfQ7>gK$Ec1eMevS7uA9UxoSyK5tzi4E_F*26_< zHnRu$XFa%G2BoIfH?pIJ?Wo zb2J0`7b%}N05;r8X26y>um#tJvDYXV`()%y>+V9-d17gEZoXw)?R#A9NgF%Pksql{ z?nZiArAfb+9k2wO);xqxS^zN`;@&HbJsf*dME!|F?t`{#GwQj5yQvvSYRP(2tG8>i;TU8rZGscmGBfD{SKOo`m%av~^xefezd+5AKyFm2yoA}DOgy6(4ZU$ybn2HTajG$l?O_9! zjRc8gC08FtiXg+^9-LJ~2+VHHwS0dmzongX_kw}X^}>A5(CZH~szT^@veS@~!6p|b z$9T93p;_^F3@pGdW4LK;K!$kJTN*4~h>EalX$Sb*NL=mp@PnIXEUZ3u&!mF}-;Qa0 z0)@kJ?Xqa*bN~M@>_7AWZ^7MyacS#B1DL9sRp{)5+0TBr9FD zaTtF_O0efoqNOv)!97Be_B4wsTgy^_cN(FbcXvM3Re7WoMIM;r42pU}2 z=N0bBA;`sf3zB2Vu-o@HjKb_mIvZ(AuD;cap6;)HQ27xzz2aIjd*ITXZ;6t7nk+GX z`i7G^yYz)PKIM-T;or`?5Bz!=k-i)(kvEB1=hswBa2x0Vj@(D>H>aUlx-8sfR4t{d zO>RNdB~0R+>@WftTZ?@iX^1--Er#%!qOx_nR0)0Lo;z=;adW zffkQ<@qcc|ODC-&t&o@!RLfos>DH;p!Y>#PJml(3{dpxLtBZE$N70uAb8esU6o34% z*xZJ?7EuFDw2*hoK);e!2z8{t}ko95Cgk`|@tS z%-l?D?NX2})D0A-81{Z?kYhP;ldm15`N=szV*SAHZ@J40KkpNhk5E76glCbpb~2mo z(Tw}r5Skw>E>!Tdi#|U9LZqrs@v^#(!(gTTHN%@cu5qy(H&%H6ID#vqAo)X=B}s!w z8gOAdmB37~I7h5$kQ}zQof0hH=E)XWKFSbq&ljm>rES(wb-hg|0HyR%W#}p3Qz^`M z04Mc6xl&&S?!KkcpV_!>=2^j}ytlX)`~d$vK)O)cCO|;zdG(Fh%7_FE4Gn*q3jnFgR zva(qj5M{@$p1tBUA@a`IBtoyhzb>A8Ol0-SbY|Bb-Y7`JWAKkhfy_l8p%+O_#Fg^) z$g{z$8fOocU1J55P3GIx*Hj8(8t?JaTR?%bbtmr+&9E5GLJa*m@>VSy{hY-@%Teo< z#U-P^TDo5glWo*SjU^P)sy}f+N`U!sHHnAvnSZ0Ty;neA96oTlb?#_t?bFLBe|nFW zJPbOl%%UUD-YGI|DwYH4x-{3OPml8|3>6_1Zj$vs?MXe$7czGb)fzUIkJb9wkOy-` z0(1ws5yB4t+{2xWgz=qx&s|?V;OAQ5#yh@v`N|kop2w8aOO+dwMFo3sok6dL=rT}5 z8vOes*Y|A;Bx!1pYy8$@O@C!}jhN!l-k-eFeu;)eU(8y>;76MRw(g!h?+~OSU(*P~ z3xZFPk@`CbRGCVA`v3}B^^2elS@g3@R10k&>Yl@JY#;&iE6t^) z;Sh^CN(O@**jk1pY$ZP*GwouHQr^Hy+fWd~2=~K8JWqf&M;%$>Z$@E~BDs#dbsF-% zSH8AlqM#U3UwO9nEjvv#MwRv<4aDG(gO|*T7(p>yIEscfhsE~rz~3DR)j3{nFJ99X zJ9ye2roP(ZGjhZ@*+~=J8DIdxKqB50Vo;1e?F!-LU-NBj85!**+A}mEX`F(l22vUF z{3FULSyEjOQwvyeN&nbcUUtbqI72H^$TD7 z{8j3B|G8GYv0XMeV_iC#Nox%KinW*(jD!yMNsJ2+f>e9XJxo#y0 z;x4k!0{MW03o-H6Htu{NGu+k;&%2+z)A@46KKY0L>y%N`4|X@UWlS}Kh8ka+qr*HA zSIB8#W?58^SO-WmA)ARQT~wR?lGB5fs8=6nurkdbQ|_cqJ}s0jbk^A2%w@Kah>QFH z13iyEOB5*2pjs6X3=8_ade7Psf_`7TTRUHjs)&3|O6D3>I2M4&2VKLMLE?`24?RIa zVToL(&x!a9Mr11S0=`87she*~zCKrupU)kiX^~@oM->}6HI&@T{3M48LOw*{3V?G! zT-ZAwf^BY_29)r5e4mC!`dNvV$P)-1+#73*w?}%FE{}k&_Zj{TR ztDs$Ng+uK+$6Ae5_d<{zH5Khnl**ZabZe1Itcbi5Z!8k$cb7;eX%Xi(RT7llS{GV7 z;b)ikN0TdCI#Ry5yJ_E$NcB1t=@_GB<>cs12UDS&c!!mnE!`i8o)|S&?(+A2Z&ruo zm<{!v5;h4J8Q5AdCvmeVLzq8`D914fTi*NK>i`?bWbrLfCpHW$_K1Hdu+_LKX*nF4 zL2DWAD)DT1=y?-^I7Bt%Pb-*`=#zx`t-K&bM0ASC%rbf0$qAcxS2)W}_3%uHNO>_; z?o7x~)HW#sIJ5U$IQ3iT5%Ghu1Snp(x0%U`CO-b5y>aN@Gqg7QXw{B4cT&9mVWpWO z6Pl`;1ZFq-EdSB{#DvO;mUxZ<9lheS@p+Sk0-c*q{huu##E<5A6}cXat6s#)8({V` zpGUfbz+6f^@!aZmo*Qsx#jUmF(M>j?&BA$Ce^Wyr*6fY(&YMpf(B>Iv%(!(|QJ6^J z)&3-&;LdD!x$siXv9Re3ST=PeoD+Y3cPdzVRseVL-gQ$ZU|UywPB`BMRA<27XOU~I zQ(Mb7SE6KZ&z=2Y+xAiO8a!|^@MKUrDfw~9sM6QMLeZvz5|?tM_1ZzR8Q$EC;M zT^GS9rkJ1qF;E;wRsA2P&N8maH+=h}L_)e-nSeBiAUQxO5kaJDNF&lMg8@qSCUNNmQOZ0z@3{{Nrn`8+T7Za3V!?<>yhJdg8x1U@0em-cTeOb`Ab z53j!5%=X(>I15&79qqL!G}zQ(4-r=)t7k&FBH5P>qWyzrHR*J$Q0wo~*KiK1!dKgB z=P3-U_%!5Jg#t-O^;unxkSdV(GKl61mEKSKG1gX*8F(pqv)3oE_=xgjKoDm{=FlXx zBYD4m$7;q(8qf0_(f7UPol;M}805SN{qh+S)W@&IK`%ygIJ5el!pOd@kRU3U1M*Wu z-jP?o9+(yO|5*3y`LW)IIz+s?-agG{zz9)?w*MN7gyX<@Yc*r6qN^eYXv%+%ZLoX{ zW4PaiX&efz93d}kk{EIR`E@((QPLuqA_(yF_A*puc z_I$S^lbu&`^jLGx+)7)4)v@F|TZrP_%#gENoSVoY6_5sjXTdtvM$uekJ!0#kYTi8o zBZ6~FOV{w19bfqiPinlKq@~a5C+)P~$MNQ}F0YWu;2)*NEtH@SUgLl+RfUzscSC#t zFH&2?#M9g&SKuqaU$!B&EO6~d((OL|{Fi%F=3gPMMlMg|uKZmF!%6J4I8H2X0mjfJ z%!_%tIAMn)H1-Z}d?VMCt7ScN?TG#4x&E3Sh*}@U>%!Dv-WC5eNiZ&Q0M7fNmH}=L z5HgWzS2UfaO@3vV006bxoZKs8l-5sXaYfmhI8Qn~OTE_b1R&u{%!yjX7#-4MVMzo0 zIQpdDF$zq7c`EKL$$7o$D>6#+cB@r)v&xJv)pvcGxSAcLl|gs_uCkAt-(Vwt--7cJ z=j0SjXAzps%lC>e*X%nDOfwV(3yeO#SgmwGFD4)Iz!lh5G=Ty92|zPt<7Nk>i4{1E z%DUrq{NcnOZHeJY{oMn0nBSRa)gm@XBGp=(Hrj(^t071|(ap7p<#-}(bUzw_9sB&zL82iXo&KqBql6?mm z`@}A3kHsNGiBRMrij|zGl1%?hpjs@h3I^zt1r#u)02YWHag`{H^P3P6!0F`0#^P@c z?jHv>)Ax6G8ed4ryqD@KKh}B%Dho#5om`{zxN(rEfB{bI=`$^L+{Bm&HEy8L6MI*? zas6vgsJ*j+^thM4>5O!j_wxG>Eq^f%K`nP~EN~u(By$=OvOvRQ08`wO)sBqZ9KY}1IW zG2(MpLa}A&gISUW(2Uw7RjzI-q#SPBshmES@rs##6d!0s^XKlDHH@yTG8OIaK%q6` zvl!cVpl-rD^d1TNMz^=&k0?AZlb3834>77O!IXxY?w9Fb+})IV-~H$b*b4~JqoJk2 z@q!7R*o(00T4+%a#RMi?w=fmTmizTR)pMnpfqjaQe0GOfhTFA{o}RDmH8P+L6g<)> z%rnr%I$8h`M*OP7No*S25Sn#gY;KiE zKEBplG&|>bb;V7B!&KpMGZhWpji=#lWr#p9Rv#1p`eF}Vra=)6_oXM#KU~EIzhMxq zJ(&%h&!@y+r7^KlX9;Q0UlSrYtZ75=L5=mKA*k$xqp#G zaiCC@`EM6HDJ8_jnwp8I=EFtA?Pw}Gy}m_F?S&aY8YmHo9MYkvAsFzs*<>@kjpI3KQMRZD zv1e8UeaChlt@j7~M3q|GC`du7zx)xg4KSu3D`_Khk6>E}?UBHT3_Xa|%t*<*d4l-q zCX25s2{V1o7U9Aj6T2?&3}A7q62?Z);A4^>me0Q)aqR zV|(e5#@Mh3BSq0mh@lsXQ~q_Z5`h#lha_RB8{OB_)jb=qm2PCbXJ*3gX#kZ;3TJ+n ze3q7)Y6yadSHif^TjfZ8EdvrgFgIeY-1Pn9qJ%U1byD_@Wmn2W<54k(32i8cRhOvG zIF|aQzAZl_KB{;SNtBrw2URVy2QDxX5sFb@>B+-7Th5pke4BM%T^kSdvg5fP-CB38 z{jrke(|BF!Aion)VVS=br{!SQ6^V8&CFxv^5**u5%Q2nng0b!s9BRhi**1i&-oNOW z0QMC(51(P0r}76+6X0_vr?a{uzW97BD_I)(h@wb*DvRzyW4PzquE072fAk-B=+8yR z;8r93$Af~u!QbBoZ>1BE5o-!5d_=lOnYUP;gqPK|UpSx+9=a-e>(eQe972qV#=CEA zW)1NB{_(vVj?EGwnI&U&9j;dQB~{r3^%v4+Za;nSt8Lc~3?wM*eiq4nP9^2$>F-Y$ z##!4Tlme<#6$CReH%htFnOx~V%TE~+OqkchpjA6S%fWj@!_6fGcxF=j3Y z8~d*=79Nhc@=Nw+8;LPPOfOxf$)R-n*neHQRR-2MPdoc-{@#gNkE^V5SQqYFoMDN^ z_V`q?XlxF;ob5uXw}2dj0ra|Zhl75NUC7zk6K%#Do_$?eGhykjTivWS3X#P_ zpPA5l&;$24!>I3$D&VgaV&2f`d&lyccF(_i$>NVO&Y=1z9Ph#W$rODi>IP-s^!p*l z_%8N-Aj;HkeN&`aPZHl+TUz+Lk~I_UKa*ogc?(X_V|&?J+L-U!UXgu0bBE2X#XP2j z+|eiQW;IoekM3LcL75Xu$Tx9tGIt%;vn``5`gdTm2m_ByDBF{;Q;ir$v~17W?fr*r8UmCDZ;Sx}wcY}L>F zO+!Ah%*($K_1&b;{8|ZQR}yFbP`(be`(5`%iSOgrlY|$Yua%~}E`kg>>cP?&U&MtCoi9@n4ge<+HMw^v~e_2?{MWYOiH(sDgP9(IF8d<~pl6)AHr zQp7skiX6;lw%?54J5d@|)ZqwnqN`muJ6KDp6^9CP^(JW9bwz%Ygpfme{d#ywjOe%m zPgMTkE5j$jIP?ww;JdRlpI2SZC)+Bigs6;pyze{Ray#p)n%EbZafiZmjbv)_gmXBK z?Ka7tho{^I6u6SZ-7hLFrx9A6%{AuiXf+L>vIrB;JkpMC}eMQ4KZ z7FY@oRj9~B$}AFR1K6%CI%>ov%uE8}5(ZOEykqrxF7LP0(UJT`ze~&tGJFAIy8H!& z7JZO-6Y(p*jE4T5tTV_+k~qSexCbJ`v%WKJ2kW?*Aoep>ewnJ#qa*$`}=j?Yd_`b-x< zIRqh@-ZbAT5|B7^*35E_)64-ygdpDYsLUOUPDMGIc?Y2QD^o}s2B<4Vo_|9%CD!UM zC|(nKS`Q&U7@;tJ!h6l!?&>SWM8?m`?PeRPo_LO07TC}|3K6Vj91Zgau^v752`A|u z$kO!3;%E`-T|_^OzbY2V@54h8dl_r@Uc-QU1M!mGf2UV-p+R;h*AQ}hs_lIYj0p8U zObd>r<%t0Ag|SJky;2vZLj$=s(2pottV8b=%V@d0wIB3;~kzH6Q1 z!=ZaRQ$ncRyQQEmD#$H%DPPXxp(R*x8s2xGeqwWB+~V|NHbWi?0i*q%5>m`q>to)hY3dHR5kIm|7CU0Ckx4T8B93YdIesa8ZIz#6CVgA{Dq7758@SJGX z`F7yEnX%F*hhP1MpE;89Kghb0lrocOC9r(j5TwdVU|PRX^}gy!ZE@qegNvgv%0TOm zhKXR+85FmI=-RUM@+-j8TT*Bd$j|~i}yMJ9lN?T!XwX7eD z4%X460DW^O`p^z>1q&DW>;#bDK%=*Jv*MtcV4jjL{-fjZN7SU!`1Zp@P_X1d4qZgR zu^_7OGnJ#(w%qbvX)UbJXhL4eyc5m#L|p#SR>9puZ9`)|+P#x!DW`UA83Pc?5Ug|Q zU~`)Rw9S&t2Pzz6yn3>?Z@*TP*igfoOZi(gZ!k)+Mc)>axtqqyrArBYpTg9C`_;~ez0ox1 ziP*JacT96N$~Ej8K6dX*<(K4_gq-^iLcG|UMfR)nLkl#A9R}=c`e@NL?B5>SEixGB zic0FoIEtC@jS z=^>dyL7@p*qVb|HdJjk>wP~zi4wY{tZ={GsH;HcN)zjf`S*0HRjH->PJ6GiuBkzU; z0ldrh+r80Or@wc4mmJC#{^U>2O{wG?dqz8bc?`H_L^f#Bei z&HvohgbtdO{z9CVb6jQ#0AMI`zyp-zfw&GVKrRO)@ zSB3CT+XyV0ZMAFu?tr$3uWEx8qFTs=;KOTm!_dXSjN}3M*shXNW{oJhj zWa-!TW#^Af*U+g^P98)`cZ#}R=7NLFX2 zJ*qIkRHV9CGQZ$69)j5p!j> zt~eovK)|C9G`G!@(JY|jc+uI&wqr?cgBTD?cVpAkSJvkz>^MT~(L(Z_aef#=H(qU`*K1n;<9UuX<9g?*NW1MUt`HX!{@X^7PUc)?UHU4P zg6_YjK)_0ucv=vk;D30r8!ud9e&IaIUq@74Y=Z9TM9BN@2W5>DLR954zhz7pny+N3 z&fCuIIhJaEtt#iY_~Man&74{HQd$o*Og=MEO8EjWF24kB{Uv2;fDiPFW5DC?4M!eK zZLF;?+l~6y_i!6DK+L(5FzFueY2=)Ns$-8T)da@k7d-ogNh zB#(SI7OKc8_ssIkA8moeh!%t@na$)FyrWsqrvRI`Fxt(1CYB%WyhWY5Vx_*5EQyAY zZ80U0LXtU24o{Nvz;?T8_ zAPSN=C?DV*YjGEs$f9*0ChXh&ob~stDh>_qbMY2&|Fx10jCZ0cTIu@$frw7L2lyK; zy#}~>gUfzzq?)Z|*PJQzXilp&@-lEZb-QKq!DMIYJL^PFs))|NmBfWimvVclttQ`U zvF%Vk<0h%Lx4rzEzkBYU?2Bd}Hjaoxr;sY4@!kgH&;?}AaZ*nyxJ9Uw*^3oEgP!%h z!2CVqi|pG+(ebKl0e?GWenoBzFOfNqV4uA)Ya0^11Nj&~Gf}d5Lv6Cai#M#2E4s3} z96M%&{PQcf4mct#=rOzpW!r(gKCrHd;qPUVwy(O2z*1;kotb}Cf$q~<(0l&9LX@iK z-8BvwlKnVD7&+68REKx+RDoySo4-Br5Q#ba0`K3O#zPUWMyK91G8f23&PM)KR{E*6 zK463Cja$RRYd3{q|3Z$yLDtmbO=c~plb)&;UqAlWl=ofPb}c)Fk*UiyMHigT+h|^s zVm=d^3(-C%v@QK^W2C&>`xrpH-*Ndzem< zM)rnBYSs)q&7U|K7o9xSmO1%Lj3L3}-ida@hs{$RQdT73jKnywehJsM-M{NGUz3zE zX(N#z7SOn_+H?BWkZQGJBuyR!MyPY)q%s&*w5U8Qfc?#YuNv}ctS;=?@v^AXBEpoU zL{?ZL8^p@jG{A{5w~At(tWC*w@Nvo?jTX5@%bGxcM7LEnVc@KvvBui+=Jos`X`$x48_)6* zfEE?39zo>(?aTyn6VZ;CwGHDTKAb2QTSS;Nt?z3Te);voELi2}hbE6WG=B!-8V0aC zLbiV_46O*06Cl-c2ipoXth6XmPA$WW^JI??#q(*mGQMP~yyZ*Nc_*2BH6tP-lj`&% zuRp}HhnTlb0JF<@?Q&G&>D#C^U#kAUOy{oZb}~6H(_Y&R?bM8M7qZ=zOaOsx8*9{( z{!*B>!|wRWFk!<-BHh+DYtBHdfnI!b=~>&?MZry?JpESGe!Kk)FS3rn>k(ce~x!twbWEPvHgt1S%5|k$zgv14PT2jk%_#Tm_k|UNZZ2IT+8TGRXSg)1JtpK;6GbUP}Hdj_L zy$5idDZvVNJOaR7U7$!6Me0FP|ucv`9E(h(n{$a>e<)tD1M2; zc#WhEu-Iv_IP{o5hzu8t@~-R=JgY;m1^j6Jowi~BeEL9oRLYeKQZ^V*^^5-SPRn;4 z{?ihQR8-1HZ^yTl@-m=X!fx!Ixg@JR5%9&CFQobGqNuA&Q%d;s=xQXQ6wl>23nSPy3Wh zZ>uCcit{?`r~B4*e+=UB4bGt>raOAF2!|~us%O>T-?{DqE&QNp zJU*XJeuIn5K-|H!aeRC5CnX$kGA!qh?iOue`Rzp{jp*!+wb*ng^vF#o*Sh}=Q3NYd z^3zVw#|`9gwck3*s5`s&w8!|!oE*E&r`eB|;n#8T+hR>|$RTr>XvJW*Ps_Hg1AA97 z-4EI46F8`Acl5u!v!{uIBwo(T{7A13rhg#bJZ3+F!f%%>0l0|GxhQ9i0^bRL1g=Ojyh3h z1vRDiU=p+0Hi+^M=KWjru$u?O_Y-8O+(k(te9tSzP)M7t28F-O%Gxc57#Kr!kW26K#Ph!a6><3UGDe^vJPOP zPY;=3OkupO=+hd=yd(TasW~b5zGL+BGm;ndYm`-wS6_d5k2F>~o&CiHIr*oOl=*VJ z;!fxx!bnm+Cpt}dL+rPjrCOArjON+vf0R?X_&ru;1X8!xo~7NLmV*4u;US8H7pQPr zGN1`u>f(5#%(`P?_Mb*FuGKGBTTEX7frONn<=bnZOb}m3`#(99H>aKbaO@jDLNgKY zW17~T1Kln+yjJS0ASnE8V6O< z#tzjZc%gngWRD-3%UU@oQqW~PDcUu0*p0hJxPzWkL8kTccHVPX2-M;!34d9Bn_6{5 z<~VBB>V+6$87jnb4ibwjU$7m9bsqn7cSW35BO>C2h&=wcW*lRp<^8b>EtYhB#>wAk z&4fAKAmBrT)1)xn*{b0E9!i$StL8sn z-}H^RX5Cf3d)C1!+c)d=m|y|DZ+O<*z1I2c6j9)3rb-bZ?tdpt;#g{U9wK^5G@aMi z-}zVDVICgLU3;@!Om!0_y|Y~e3DIp^0!h-%!yG)402iV8Na2|Bbq~gx_pGfs&rmzP zG=*KdE{&q$vr#^B&sw&iqEY|LNq~Y$kWHCGawC0389{NZw0_yoK{Cy3tU0~sw!;)n zwVcL{9O`FKK0et74QGF3JKQ4$v}c+%PDOw!89XeR_sD zeaKiE|LTQ!{gppPgB3|I_9O6aivi&hgB$xf)4Wem>rwj4B2m6EjQzMa#lc=|*WYkC z=O|Y>suAd*$=o!ATHM5hs4&rdNTVz{>p2l=|7KpUBk;%US;(yAL&n? zm6!Q4HwG6*VXR$)L5_cdd-i@nL|N4fi0BU)JkSW=B?XhiPD=+QqkyCMux69C^vbc3 zv*qr-pWRVykx8pWydjV4WMd-p4#cjVHXb4-+D%_KUxQ?dQz!c3c}F&s7s;>|qTlT< z^!%}8x=+cXuJxQD{qtNs@k9pN5C(1%5^Q;R@V*i#AK@Y}pX)?>yv@Jvq)v&&dw~;-0ZdSRM{`Pdr_{#z&C!)qIuGDr!6tDnRm?lI!Z4Z8`We*6oy!wZV ztbM}#hds@5?Iy~2E_LSZJDv+%iZb~q7A1J92pzb;R?5=1ZEddaE8`W-zB!Q`PWDrE ztNItyKA(XrbJ<$zZ6yXp4+&Y&|28v-V?YjFgKbrAb}0;^bABRZh}O}b&wXE+Wi@wz zQq=l%sN8~^rQN$`E~)cYQv5rp5Xi=9JQz-uFcF>0_cl3H(CIUttp=Dt<~3sCqVeI} z`;lekhNMhsuIfbH5UWC&alJvv=}a4_;}T^zA0e!<=7~DY3tHMWPWMzxYz zbVX;fRf8%EEhhpVcqFXZ?rQ&Y6{;dkm;AuSn6vpV=9!%jOA34KWmoVtb)kbtENq%1 zcvEZXmo-TU;f9x-m|1Qc66GetnHBvsKJ|Z(Bn0nh-tfdUai=CrCGjvK+eCsb7V#0j7T?edzRzd~8M**n-nOQDx}$3G1x zSwN zh)6-#qN|F}ZXXWOLPnKn<9X0p^knWx!pu+Z($Pg`u0$h(+2cuJ^e21;aDdXTAdrZ* zc!?2$$7!)!KkogHY0NR}RPDaB+|xIOJ1-nGhLq!P7+<$UG}EtZ1qJ^Esq!4mpB#zP zb_7jax#$2%r&#XOr-hdL--Xzie8rVDmE)WGJ{cNMD<@1*qw90xDcTEhvSdD-(de5S zXAj!6%&P78IY+46xk8*qJCrLCWtC5v3CAw48WJ`C5Kgu%Xdh-G9f%d6#|9I~9yLNx zBc?rzvNbQYS`d+8NP6>%xA1w-g52)o#_U^@Ro;WN~1p$;mc+?N>ylhr1UlB+YUEopxOG3^-L(cT!^qa_vtWM5kLiaF};s|#oBNN#@wy` ztfys`Xld6nD{LYZG?*>Z4U~4^O%tuEY9RVnI6ZOt-Z&-rh{7mqvM7fZY%zo@0q zu)Ohk(_E-sySg;*QRW$A%d=M}#kR8b#JNvD&PShE(C?~|IoyZKrEqlzrGc#JVNF>Q z&e!Y*UiF&RXIh&0U(n~UXr}QGX|~Fn%9tr3IB!WX>;Ai|ydX`MCpE zHD2D5@HO^P7N-8(j>be|8fZd3JoTSg42YE^59x6VKz)hv9&TFI<}{tV?_V1Y?MgrL zYY51Xmi4kZTd&y2D?If<*X1FF;SVVesEIIGK-lxfcJ(}1i9(JbV;lj3b5S?fd5H`Ea-?3f105yZlQ6zRe|AwG9yq+`*+!>^U{-6t=Fg(join; z0WF4Tjh6=0G{Ae0;qO<9lMnNEI5|3^>xn2@S+yuXGL61yEtu!Yah^6|x?U@nH4ly) zvPa*~$RK4j!N8>M`DuJq^u5MW!|>4;ce;=5)F?MJ}5`nl-m`EAiO9i6!MIt;`H zEb&?IVt1;%_$$)YWzK2eP`UOzTOwkJEu<<*R+o<&WcezVB{54m-EAvyOqc^aMd8S- zH%+Cf4>?=SSVSnkJ)@F&V@ujFbF@7E0rm}&!l<{!Dm+ATN9Zm{;!cU(pLTm*bv|~~ zGztyyugWwLY&`T_%k1Y&wydLNgSB23xsm%I{OIg2ic!wZum;_k7l+v}&{KLDjx(Q0 zf^oE=y=GR@WWj23nV2);Tl;O&|Eb{gXd9vYZ#q zL5$R`Fkgu8b@P}1l01uIDAc=j-mZe`h2c1@2b&H8Ec~HFrios74{Q7tJ}u3-#g*$n z*Vo_rS2NJP>_uG`@09-}SZz(4SLCXZD|52Pen@5I`XJ&zh|c;YYX3n9D`<({<|XtJ z1Gf*w3FXA+C2c8zn3x%k{oXuIe~F8|uV15lJoUK}I}G2xHL?hX2gS{ZGE&4hz%fB- z0OsIGbR)0%Ih19+e}we*MC}UeU}uk@X}9)k$;D}e3g{2ZfWU5tl4 zlG0+>E;GOL@YcMRHV98QjPtH>MW8yT#B_=wjGKDnwc zEehyV1Zo*UCS+~CX4eUM8jELN#l9{}A)h?3-R328O%O1E(R&49jH?>~)wf91l6!Dp zBdgctF&?i_cc4E8^Ae<_8=qeote=v6N|i`O(tZk{2ROhkEkvG!u@CT)O+RWOv1Xz! zxPdC)di#CeOIn7DciQ_Gzv|~&{kEY^<`c`+Yg<7RCDCLqQ9=swJL%dA*f*P^-J5+i z-ej$%&N>+Q%60Wjda=~9Ju58W_H>OJqGh`lN8pweS!aC!p?J0EaF9)5I1~Gpl%(IH z-LR}P9b!Dkd&013$I#F(tWU>YCq``htY@LG#JLg`FcYH+Y{VJESryp1eYMCxy}1+2 zw;fI1ti9Os3{q5w;?9_CUpf7FK*4>`HX;U}SU!S)Q~eL%9MMK$oCZH_LOE`ccR(yd zZ=lrlQ`f3OhtmYj6xaCK(GWUMT2DKOk%iLu*_lT09{I=&+s>NXy_pEx`md2iC0Zha zsHvmfgzZeHnP@*~@EKR-cX`G^8gx^KE`8wa!hVWI`oI8Jn1(Dzw2t(w@_pb^;je`< z@jbV;q`Y3A+sU7oC=KDZHoOHm$aKf(ttTMIFF65QF~a7gQG^kWeVw1gI%X?udR`^3 zXxiV63J^=ET4j6`9;|bamup2eRLQ!MtEgeSJwfI%AELw} ztm@B4GstzdSH0A|(GbF`#heQUH$kHG`xpetgY57gX%qbp@0uAo8-`}tbb8a9`s;O{ zZV$qUuI62+nK;KwCLx~>CSU1WWmbThR2&uluiy`o7P@vtW#S++Es1NkhSr(aCK)fow!L z31(*A`9iHYapU0ed3wQWYGaB8!{7LE1?}}d1ZUeOYM$5< zF! zidh3iMT#{*N<^S>Hi4_E8w9~8{d#py2AoYQay6^c#*NYX*4FXjiB`!I*ER+yz;tb^ z64-OIIw6$jDO`fW=Mc`A40E>iX(&^)Sj?Pm0^-XGyK4UHfL&Tl$wvaQ@F6>X zgpdS`&+(RoY5TeEQke$vsKX20MJtYW@7}(>%Luyto0}QF>hXz@pKWgvrS}da5P@xy zDOb!ijpnce<(hub`IbF9HzVnK&nKpR5Nf+^NhsuMCCkzc&?$16BCJ#on*4_4>Y38C zd{4V~PM?ssYnzpZ10H$r{TR6{2*)kGCZISva#Q zuBG|Wl=w>sbF0$#a>$gH7rN<~C|P%SZW{?w6mjMv2*d2a z)>EM(S;KSCID65vL;!cC){=?Fypi~Rr)RGF>zCJlJ|*z;7kf*f(bUw|riISyYTUxc z^aMt<26C%xxVr41+V>i>!8_VmP#QG zh*+ZD?}}A`6w(Gx2p2(O4KA+=xi zm?8IhXiwMTqF`k1hlKgiS1``c(x@l_vpNW2_wwN3=2AB#_88oJ3(hv<_n{ibHZ@F6 z;Jd|q*z%Q_n5K-V5`#p?yn7#I4ry}~4bIa)TQPF7-Q6Q|j$r;|2g2?vq!cTD<=yn2 zm#^u6RoJp)l}MYk@bm@s*$>{C&y#w zH2iN-km85z!~Rvh=a|{wzod(Jin8m3RMJdc!64hg?pzF%M$brwfH2iI>a0KEIICrP zd3l;+`Pw~vdsemm?Ja9{%f~}+bItjOAfmJBKb{voPLZl>qpkp5)#?>E2T^L`h$XRi z3(oDka`DRs>O{t#sb0+2jhB6GT72|feD08k4dgd%-cuo@VK{Lb+@{KDFX)DUAE%A0 zNYDXprM|)NZ!Hb$YcuzYE#Er$p10GVWRzAYvdjC)p*sos6n^$n7|j3e&@kN61IbiD z-8W{J4PPvH^!K6Yv!!SkXPED<4?g&ww7xXVNM-?3G8U*C(LV5OSeIGK zFgo3~M`IE2FI|jVi-K3o{Nc6G`ua6s) z9Bz}u(1#ami3+1d8)~y;L0Gb9U)Or8DZx*D?w5G31XXI`O`3cDBh(Y(0jyK-c1bel z+)vEX;kin(@d8CA>dt5y#_mXgV`Suy7h&Hf>ivL+Wy6=zfe6#A&;Ct0r5Ae1JOl#f zO5md&85zEt0;9=d6JEH*MxnNw@y7J?7tq=OQ?RtOi0)|{vF<204Gs>C6nd^IGXQJ9 zH73aS7FG3d;s%6tnUDv5yYIT6s1Vr~Yj$~?_hDE}EYh;?Q`|($=4m#8CTvC>OUSE(g?b5y|nR9~R6}Virbx6sFrT;qSbt2|i7M&i275}_%asBxk)H0X(F*|A} zo`$0`IG;Qgx1f=<{^w8FG&sd#3-yqm-cf44;DY2A-FeNN{0@b*enV7zu= z(5b*M=KG5eY-Np3QgrdONe~nlARqIRIX4NcFF|Kh2R!{%&S)D;RP$E?#|U>_%F)2$ zHRp#Sq~!AvQWmR0EW1alTa@8@qC9cjUza+4OVF8Jj)K$fJ$Bv;{?CiRDPASM;Ut~n zirhlkMh3@yer3?C#$iCWSlCcqpKa}Fof2gW!4wD4(~9V78yfA|rUwRev$E|MNB-1) z|L#vQhZagHT6lfR%PquINgav`1+)e4!yJ-!+hH0GLKGYpxzOEFV*1+u&aF3-FFT(+ zv$bfR4UkZ?7i4}SwiR!vN6y5-$m}`WIu+Pck^dlL<*FZ_%^n6|OTI40Hu_aat1(|U zg|d&^CB&ODduwfT{{h_xXO2K-RySSqNx5gwQn%E$Z3tFn)U2HfV=D_#bEsR&W^uJ2 z>9^?*!BR*25$K^!$@lWHd+`YC+HXzu%d4yF4*RAWkH2)^7U(i(H0iU4K-m~yT!jkB zZyq>;OV;o?2j642-`is8)9;7#BjC@>F~sNd9G2~R5!TgD$~~^VxgllEXv+`%p+sVE z@yAE(#wa9-sPuOhw$wc|Sn$ZU`&?C3)0O0G=bCsHgb;k#cvy-59RP1__B~nOC_xKp zLBxAsM(Ms>0HU4lbgVI=he3>{U@a)*^0?ohKZN@@o`)nQ%KHH}9M2nUjLhqE)Uzed z4kngNCZLMTbKu;R`!ya_3K#mq90cyZ?7 zAb-Q@25SGL@v5&=gRy*|CHLK@>;jN_R>m{DG{A_VC=8aic@AM#FOiQy^Y{vM|Ela5 z@~Ycf>tL=S z1NT12ZQOo6-Gp8>#Rm12?XtN5>i#QAfL=rl%Oa50I_Uf9Rpj&wmJX=ClYlB6_Ro zIJLwmRf1UM^-9y_D@Sb4;`r4I1u{kcNnh4OOeNn^H)1+Qb22@^q9GmkQN;U*?2maU z+NAyU>Pc*=X#F=AAkd`Me)A?KCzD4ZSlnUs6|{*3>&~J6h8YtYKPpAL4TUgl+5VzE zOC7Iei`jjMfiLfE9Y4}yz$&n>>fOa;FqvIcIR277D~URrn5Z?r!rz19QW67xtq~g{ zKn4ni@sfWX(R_5pr8`{6%O?FTf$0g4qJFWAU^^oNq)_o|{(hl3tnZRpucsz;d^yKgxt44JbnRy@Y;u zSsd}&s6#~W#TVzKKd+C3xHB8mJ-WH7C$_B}>>P`f0$2rY{Ug>@QG=dD?_`-b3Rd>k ze2s6LsDgy3si+FrSb}u5laRLSQvXw-0SJkg$gL+ETBkJ>6%)Mopo{c;*e7&HB2$yI z@y!}&4_xl18FNU0jRgZU6|Im58*d;1WSbKGiLVI(O9)@Af(Ws3;&#R83D*g;l5d(g zG|2gtR;T8>yu1sKbHx~qWKFghPaix7b1uu2a<&dXrqFuF6hDzt(`oOd2i^3`HcAyO zEaRYO9E+CbkJmaQ=7kOcOtHjJ$_Ef9isK$^-xhc?p1_QasnJh^{<$+1DyAE9{oWk6 zWi+T{aCXHC8bDO2(!;?McX8bmpqv{YR@aKI6!(S*TYeP1>He-w{j^>F=m6#!n4uP- zJVlF-^0vAJy?OiNX&Fw>i9P2YMT%;Bo=gHXrM>Y>e!J$M=dee|gxf^3&}fIZk)aQ4TD+sH8#PNOy|YdI^SooC8*khViWI_U`w`ntn_G|C(Dz^& zs10-I56Sf3FB9^XCym=KW^r+tP+9BOe+2@XGct#5%RldYx(gn@3uV?~b;C?3_`87B zy#{eeLujxEit6ED-mN8J{ty-;VCs0I8<4tgUHGaNXXIXg)d!YHWl; zQMVvM`EzJZ^>{Hm*W@2%rL$Wr< za-gUXlbsR>x#2Vg)GVGy5s)@`qg}M`y*GyC(9vqvN(ER}&kb zLFs6VCUjYlHsR(OaK;k{B=+#$4}#6hzLJwKUy3Z9*H>&VJ^?p$%Vb9TD-Ed?WTZ+*ke}hW0szBa zLAw2QL9X@_DbqXKI^JbvrAi+Twm14>FJq5~WA9PU$r$2W(~vP?L+2AWZ}$}03;nXX zo?9gOk$LH-kzeyZ2smjo`g%?O7M;hRbtw4h7?|t7=+3gX?fr!G{tZ*;B3cdJ*kZ?Qso7mGop(S9|wsM}R^(^jJK4s6(W&NhV zw=Fs>BRCcIeN)KC=1;_t8;bvmrd6IW3%@<{6Xvr+^tpcM&evTFvRml+J=0Rh0{aU^ z60AabFW+`_pcr$3h0=p;8KPj3NnJ#=r}Aa;C<NfR<%437aZ(VktNQt{Z z*1`L)`fhOIb&8wU4U-L%|Lrf-mmuwSWvLd%wA(Zn&@2F;rBSwHv_*K_@v=Lg6k(GQ z%-)rHzfh=tQ1;nl>S(G|A~@q;N!xzDk>YHIR@rB@ z*oWRjVwXv`kV9A6v`p~dL8HUmfI`$Y!(O~JXn5?8{1@~(D&y2S_c#=tlZqXQNuUMLyhqM9_4XYJ;`%o3J^{A?_`l){W@j>c{S^5k!Tu4#SzjL@D#4{KeQAkP+wheT7`Tf43XhRE|!4dGlf=i*p!t zhPf_x=tka@DM(4&T1v5z>t`I8u-XC8P!_B_g4S zK&+}Ywo)^4QV!t@f zvYBnUYYzQkBLVMjW<^eO?I?roKzPYJHc_slo@iz6(zH3$}@qHN})I5Wlg9srzrV!WWA_`8mPJdnfWaCoY# zW$N07XUQKD)2~6J82%$TA8>N+Bo9 z#L2lYsLq6%G#LV>_?~JQXKQCiLEoby<&pRsET%5)(m2x{IU71>VoanYob~6gCw7e= z!WYH4w=DLdpAs4zdVRP|^+&i&zXcpP)7kf9Vge{WRPC$B^O$rih@%bZA{GgzBUZi2 zmiAF{iBBN(p^eWv`omJnDi4YSB>dF(6|B^7j{EW%{QO6~n*`$^O!6fRm(HQWYCbvl z1E&g`2e0ftJ=;LcUm6X<2l$BHTCO~y_Ny!Cu_WH-#%lxZyc+3TQ>eWgkXBlUhYD4r zGk>10d3--~=t)77VW;E~CjoB|XLq+uT%&v(5~WZuzaIdWG;4I`{Rdbc@QW)YBb5ZU9@Wse${(0H0n}ATgaW86T)9R?z=jb ze$buoxbbnwCFq%EL+(zN+KZP1aqX&~K0>7FlUWbF{zqu{Ut+t-|AAT#1Ob152qMi0 zJlX5xhNKKK3YNWx-f1tlZEKBa&0E!HWZ%*Q{o&C<(;Lhw(gG&-0bBc@o^2NGvewZj zU7;x^LR9fCg71T0`t{dm!X-Lu(No9yi^3dKfBe2M^@>-=S=#4r%Xl~#rLMx_F(Pzj zYr=vD*%!xcHyF`j@a@W>mrZNT==GR@ zmYVo3mko4^>wXF#j@uN!>U`(e7b7a5(KO1(uydJfd&{2x&VcvkEqMqMPu?eS(DIP- zlxtMTo%4ebYH=y`N%!yiZ~c5UV6n>(J&@pef(V~PNNr<^cTWN)c_*W z;64>6H3z+jD^7%hSznVi3Q-l%s1R_~&Jh^RnNZ#6P=SxZZGZd#RHNYykSQQ!&jcR6*r~VNc){MMBPor#*l7 zRwMQZbzaN8^Yn_WTlF+s!AC3mmzVhqF6F?UN3h<@(Vy5JM2VE0qGh7|>jVd}L%AA# zD1nuhmf5i#m#-p`5!2T{il10ddzNc;nyZGj1%h&by*bP_;Xs(u%9~@I*P+Qvz?!V} z$;Ry0JZTr{jd_P(K~K4Zda_KbMXguEfzyP~T3QOuvjI1LX~3Xa zxWU6i=W&)&#lg^*nXoEh12>nyHG?6vFnV0vQxO!#vrw7Ar zv&@blf#IV}PI5+L=IbvUZ^XEVOm8FYQN8xCo0#k}U6AmhAsHptBR-rKX6$_N%~kd0 zS0{oejhkgp3!5>vXC^Ym^>|;)ILLMZpX3~U0?bL02sp7k6Db>|;zX!_h~=Mo7JFsd zP2`{V$Z_gZ`LCmlY5JED4qNV-fTnuOL7IS;uG~#_-h*Zht_GnMW zD{I@q+4!iEy*`srFMY{1d$*JVLw;P9B3KUErMyCuZPhoLrowEb=0Bglab|EHCnc`U z*Kl*Kv^UQs`x~?Q24_|1zsKhjgW7wO7%m3Iv-HjJ^@4?7q2F_F31`Gi zSDuhV=}R~Nt#w9glktqpj|VM$GD5EM-na$6S(`oqWx?E1Gw8oS?!~1Vne!CMjrx1% z-SSQOy8Xe$Lb{#0{)T6{O6Bs#gT_sS%v&Ax<^O^DZ*nooQ zUD5VJ^DA;H?WTcMaUoVP=LEL*C5aBXy`mE^5I3M;qJir5FyRF&i{d{%)?A!afE_W=)GcppmV>2mA^tm!NS{>p@%qE=G@9^G*CmxH9 zolRYAo~LzF&H$pyw=ZTRy{yt%46t!D-kNfpK0QvcKv>WITuVku&xO(80|sT!0>W#) z&a5oi2FBb~7*K6IlX3siN+j2|o3g2NIxQ1CuGR=S@?%2H`a(Je=N)zDu;20flaaFd zRr>L^x3?6_R02u}p-3>UPkHTr?pgpw zcjKE+YDqz6z4yZ*iD$Q$-a>dxMX5Aks4t9KN2gG2Ufz1*oX}G0I^DNnCD-X96L(SW zM)!?j#Sbq(y-)7<5`%2`Mi?en<)gB&8=k-dA zF7*pq%D&kWx_tvIj*oy(o=TMnUi77lKl+NznFjTmWIe+bK*HdUbz2Y+b)B&2&p}IV z*_a>BAG!f`IN>8gWPxh){MM72gqAwxX?oOp9-&L%HOH35Bd@dV2|M!YZcK3(k5U1t zoC{6-jan}pUq8G4>R=;rV}fkOk|nF?AbjjNTZHAQwE9=BYxscdQNs zdebwPOjGmd?f3ah?7|{Gd9tQTKctmEDTb)zNI)M#374UZuQt~*!2HfRAxVbzxDKZ^ z8^diIuLgy^I)Q#@+&&%_YNe5bv4_$83sVSLLoG2BS)sNrE4f0xM1C}-8{gNi(d?cs$7x9 zH3K4ua|hzWrDj&xeR;Elo(qWNCj8QN42&=#0;TAl6}m zTrY-82bSn5MhGmZNo+x%#J943cpSv_cFbD-^^Ybl?hLmRNm%Dw18~ATs&R)V1iBB@ z`VU}Ck?rso!tF~@M@_*ADRf>{6!{k1hwk z%a8Ze#Z_J{-NIU)z9E|smZ=>rbD8TPW8uaBub&O*`Q40)eJo_2ZuW*_O!5sgArwU4 zQbu>szTw9SlOWn=hMp)|ngA%C7?KX5zSC>NRsvVu=r0~tYs9U#@{^b*Cta&$8$#4& zW96#yZM*&*n90EZ=kkGx@;{f)=YK9A3|u~6#0^?5P#|PK!MCN+Ril)b+~@Ez6r)Jb z+oT!M_z}w(KUwV~rcS7TA-GO3489DJ6!XQ7VNOo9X^56*c*Gc~2ao|3xYBOzo zs3E873)8OU7-mi|twrHp|Lbc0sL=U9%Zdk=?pWPE7kv`SXcSyqpSHfAN|#y)XSpQZ z`erUr{if2}vFOl|@%Y^QEG4F$F+kK|-F14oW?Q>V)o4DbSBQqce>W>_`@ve~Mi$P? zX%y#lCO!@V;rO%JdyI|H)}PJswTz9RTf~!D`Gt_G7I@8F{#N&I`38_RMZ&t`-yqTPlw{XuN#(Ofky;>W+{=AZM@2na5ePo$B(0AY$*!JS)Z*-ga?BIEOOhE=8`;C(Ar|^(Mg%D>3(~0#7 zJe;CHpSc9Y8G_+zD!A-aomkjapi_T=UHRJKGl*{{EwpJEsWzGB9gomlZs)4`PuCCy z4ns?-B)OcBj}v)-!QG4sI_x_LxYsux5d5NVeu4L>rs{1x=f*k5o;Vkab!`?ld8S`o zUW|qR(Wjim#{I`|{o`y4r-$Oj_POBY;Xq%(aTUl#0Ugz0i1#{A!87-o%-b3|!K~l! zc_gH%2kq8hPqSvykiY$sT^}M414ikbC)rftn)SKrUxV*b;&hf?S^0!zd8swuJjc3? z(uh{c9d$1Wn-Kbg0KBCdUQ65*si3EbHH9 z)R)C%uh@Zss4G>^j@Z|GcHGhUoj+>1@C&SR^Bk6JNJ&@R|Y{R(eoW!rD zP@SktIQTB`d_ouzgC8Jxac{fSDz9S4UivpK{a%O^IO`*L`uBFU!nJe>@9V7z)(~NN zdMNDQ3#g*?!$j)44+>j0d#r4Ne$V-v9v+I5iW$Z<9S&qOM|m#cpA2De_=`gyrQVl$ z9|})M2DB5NshW;T-&-Iaf@Y?Y;vmKaR&5V1&Ln1Xkh6FUeqSAyG8-}k4T%^ffIbat zAk`7zxRgv2E%LGN8!V4E!((mEG~3S?X`ag+CZ!$oF4WE7)z-N+Hxkj?>$Gfgo0e%} zfb22A;F)lEL0oxn%*2r>^d*7b>Tk{>+WsemdN1JF6$ZN9v{@Wv>(|-clWtgwPy&!+ zaHtM!F_9EF*$h1bUcguDvB&rA=5~L6$;1z~41S16GD~0kaMS+UwU6y!wAyn12&{dB zcY7)4zJT1=;5e$y20Z!CUgEPaMUE@0p8og3NSfvXeP0jv_=O%dy}0b&@X@pfO#IbU zpuS6Jp~{ghi5deau7y7F?smN-w8G4udS#SM=3Zc0mXCQ(vU1)Dk53G5tVDZFQYz`o zYfTt%*e=316bIHsqx?y5uBe-^#()mBEq0YLQl7wy959mXn)l&YdN{OPT6q=<&KB$v z$(B~V2S7l0n(N1G!q|Z8$Ku!Dv02ajqMv^)(wm=gTfKNNFs9ZO)Q(-(-ci28#+v?T zhZIIDG=aI&d!WY&!# zE^4)yat_t2Q8}l&6}*VKY?^QJ0sqp@K^yOYkQa4~Y_lx(vj{IYhU9-5X{N~#6wl=F}7D+y_ zi)`z^jxu~X>VO9g;lHNv|9+wVPW?Bw1bN|NaWDKknMcbrmMFNf(64d-No+lcH5~YZ zO4>#HXqu^4-YNX98gCfI zP_91BNWONlp3OQ4N*B{yj;B#v{ou{41GcRFNLa_zsZv4D*|6`1Hdd8apC-J=;WsHG zA6yDGKUi<4NizTt70kzlsAlLOEPIWm%RGkYb;W@6tim7f$Gw~T^um%a%h z@Z$11i1GGX55~etAe%x5nzedzJ{aQ{%iFtv?S5ub8C~_w-u&)eztDR_8|Lb5_L&Z^WFqv?9x=>FNsZfDBBeNG{;h`_FY}2#j;;&_uHZ!+Rv^Yhw z7cr)1>|zlqXRO^NGoXWZN$@xD-B+$w}Ow?Z;{3goZxX_cwi;f9AM z?qfs>e|3vQb=nY}NfEtwX{C4aA0a;ASNZZfTT6WFc?|4XU-8WSz6Xhik+C;2yvjeMKU6jNjfIp|dwikpw= zIk2irWc>aI(?2e(cIkXaXFTVy(Ub>cy&M}_{1Jbgv}EC^^g1%w7F=*{T9XSMp6;4@ zEJ5WO%qr^VPxK<{(>{pgReb;Ftb5ym*!}jzetJ7 zi_+3E3#`Sm5+wgOsOek^6lL}SFH#p_^x`972Vb$B^0v+AD5xj!FBPSa(~TC8N(dGL z?||LWXkTX{%LAX7i0gLhCyfZNOy+iO0rYW21v}a+31DlE1N41Q*2!*hss;fa*QIsu zPgZV)Ut!^uBJD=Qab|vHOPO1JwSepN0|M%5ktYI7 zWU>HVr<;3LO|_Y35hZuB>fZhNFWU+65~prNXbE3nWWo0^c*YZ(=xjtUqT$AHI$iz? zTu@Gs$Qxt~xq^V(en!)5{!0VAKK$WYIb zB{nsr#%QG*t>RhLXdA98AujJ8XIaT#_1NN+yeXueT%@jdm~ zu$<6@t~Z&G9o-&3(lkW581j9at9H*Yo!)80Q13}D?>73*_3 z#R2x)4NMJ+`%58GoIonVtr#`5OxIm&eh@tt9bzHpUcvvg~ z*^4QC*zLpQE=%Iser!bT^Y@ePOXDbxQLYzO-l1h;R?>+hM&b|xaHmlmnn;B+!fq}i zoCtkXKZ7~?E8iMUjGA|p>|>uut&}5Yp{kJcC5#~qz0kOBPH+JpAU_qFoSagx@0MNh zze4ad^_<_6h?XwZgG`OoifdhAPHcG?cZW@ypLUEsiD98b>B|_1NFN8GP!ZVtf&TPT_Se{M#YspN7o zOe`nJq;8mMGQMDXe%JFnJW?Th`l&H$q^g2lv#hyXK|39S8nqm~@H7tk!{Nszr3zN+ zL&U@PN(_G?3ef-8zg{MOAQ+4@Jz8N?eh-8W_xbm~I^C+^1$Xt6BG^`K+9lCXOm~YQ z0_WxDjo53?t8I37c_|D%{J|$g?c+3p(6(>mncqk=^8Vc6A}pWFH{Gz>vVRZwIaj^; zEbT4N!h7QMrGL^?C~CAgMC?oI1zHYWBI=pz%{*e*Z&j(4HZh)Jqr&T>(gH3wxb9At zJ&!Z_@E;H7cd#H0Y-A6-MI@u^$U0nn0=L51j}8n72=F;-cD|+Su1L;E#)sW$$&yV# z1rVZXI&ZGw&A_Ul;tj2WellQ9^i+20G&us!AMNPwGV#3b+v_QdH=j=nN9k5xhCh78 zclY309ONWy{UNORFJuUltwv-uCAV)dZJe64C=K;Bnv!+52%>K1X-_3_m3KH+B$!R; z>!TNMGr-G$F+Ob9tO_~kih7`9xGxG7@Qg~jMzQ@+ICecSIM7?Fh|TBGjsX%l^y z%YQ*}SN4;ocvG^|8FjHofuk}Zrp(Y}%=L+-a?(>G25-Dx&Pd)Fnjt5iA#(j#T*0pN z^GqkRhP?MW=M(PDtZA$bHpTF248Z{v2R+hl9HrOmsY*^EI!@qT`soQiFMY8R?JOMW zhms5;zP@Q?)4(Sur~1f=>wHOqi8U=9Twq=XhIa=a|W?CmSf^xGTUw1jlzV#I>|3N)>#Nwn0PU0%3a!jGBd znAH)h^{mYd=U@GVg#o=X4RSaRF6_^iyAt5)l(}2WC##e9nK?Is(HhGs>7AcdwQyil zMei+dO!|!W@dZwCxX$}!QBzOT=a1(0YaS7>=vIDh)^mf6c`JcM<&Vq)QUM`KD>YKU zP4JMv4`NZtr@pA&cr%k#8!o9~bS1RPOQN+}L_ST|&6>65$1Sl~I+RsNZr9*3lrwr%L4V4|K$&a?d0s;>oA>x@ zwR}9|)HCb0O-2o)pc9zu5{ALQzaMP&K?x2MONti&%_ugMDEr+ZSt4~2UtiId$No_B z!yV;;*G?T$Rhq2Z(s2QYKoNYr=su2?s}A#-+2m|(e>;n5&1xWdbLj9@G6;8KfQL9o zbZ1A-`p@5%eQShsF;aR1uHfuXL#mooRxjB}773HEdW$=Sr%r2CjYlgFxjn~c0-I_c zD4oNZ;)k4C18dD&h2*;+`J$2Rb#D6Sn029{)%Anpg?PkpbeVZhgQRj2wt|PnTCoy5 zB>wb8RI}wE>{!%cKU}cOepT_!jFCokotwm4uY>FDjZfHSoC2#vX4#>TRF6PMCe;*BC6Whx{gr>cPSsz_ z3goj<9=L7#^jkgIIkQc0nt?C*H)4lF z8kd)ivhRL&Q#p6;wr~EJgSvg-m=?=yMpk0Q2fRc?lTl9t%8Wba!^YtrGtHDv z!?%H}C+hZtKwP60ejgSC0MBkOFM}CGwrHbzlis{6_xxWej zNo{GHzA|m?(;TX5vtvAqyFBu$^Uxdg##IThOVn5(bYdbq*8}uUizqVuM=zx%SnVc7 z4Zjt7LGEtSjr&_ni5Xcj_D9OGB171Pet0~J4}`Ua$;=rUSNiMCm-+8!*aZjei}{Sa zPiB(l&n*K7aWS1kgmTmRq#?p#_Tt45*;b}{OGZA}L^bMiepJf(8Xit)8|>x+(k8qP zP4H^-+i8vZu$V{HdCMMs_UpI;cz{pjiDN6Th`V4a;?dweLCn)M=7~f7Hi^^0+(v=s zGKqlRouL>Fp*XE>95JjdXapM=UQCE#t=%PheA1&Yt42qE}p zO_&iU2+o-%Dwc%31zfXTwIBBR>o;Rk8+MXNe`~S4WihQFvGbZ7dW=kGI>L~$j{Iid z;|CGoVWip-gL}&t6Nk(BsjLRB0a%aMoB@UvIp+@bwA>yuzEO4&J~Q@KXhe)T)B^^l z6Mxj=1P+`AL0yWQ!Kv6|4_<#N5X8E#4&@EFoP~U|>cyJ%er`uS;3T+_UeSt;yq;mn zpFc#MbZUv(3-6Rj^+_JD6Z2h@(^6u-{nEqQ(K*gpCXV$$DgaC}!nCG}&IAyo5lcz+ zcK)#{anoY><|Ps7KfdMfF&2Ph46dj9@!6VPgE`Gkvg;=Pj_hK3YKi34@UTaxMId1*@oYt@}hat-~j#n+eIjoEX*jV$} zg#JiSqUpnM2*M|j@a9;uY@zIE(qwSZt^3P*#l|DYxV1HDJt0#X_5(WB@ z-lMM(=9|X8aKhT4^USI7^WWT31M9=$!%xXxQFo}edErFLGnLgD85XPHeFIy6go2tV zASY2x@BW?g%HM2We-~GO7hY`{x&7L4JXorwuJ3Wz8=+EH+(Ye`+h?`}3t`Xts}bdt zW5_WB7-*FrQFUg*YrD*hw*iR}zty11sqld~$o`}9$cNBWoGC+uJiuv_;ICQF|F`qo z-|7E`yO@4|2iIkNPaIT)3)v5`-}N^$_M0bQMbPt9uz^nqkwz`K;gA+~OX^vyv}kwd zy*BF6V$wtMb%LHXAvlGKBDC3&D^1$f=EhsPU&;s_^(AxfX@V2wA@p}@Ytd0XK8mUsPvUw!#3{MM5HaaX%mFlY3|2ELGi#I|iY zeF*gq?fJYJ#ggi@_Ew8{rPacMhx_L;pub!Q2{?q)*~;p`m}@sYkRQKs)cx3}=2W_5 z{_lmaYx`STfl&#Y{3-?r^ZbgF$^S4L7A#H(y>nCd|+2O`cU zu*_i1>QcmU2H5iUv*npr`Q=*r94*qzmp@VZpyjNM>|+T(8PFj>Oiik#n&vT&VPv&q z+q5dG;%BZ}LTUGO6ol4&D7TgotMaK1NKl@1z|D@;60$_`Txp*lg(pWd$0Yx#z`iWx zEUn04AG)YQdNu%N{ce70_*9r}@GzZs#>h74n%e`Nx!V=(QI6XVBNF{d@<)~#dZq7} z0`|%@s#l)cPw}}$xtlZ{ekMdKervNNG$VJxXtP`8&pWgNfwz}8v7BSSK3pRi$ufmd zi;fXZj&5FJQ}fPKFl=oaczJ%O#8+%s@+L-<3ZCpEjqTG^yO(}+NwSO7YEi&l)hj=i zztwU5FqaT=&O6nX>+tB#<&z9E@<4lb5j@d1+WtcN=bXX2ekc2- zCHhsA#{{E;oUYchJITL*(5-eQf?tG z!#hhY1b?wi3&6Ob+wi&I`v6NVU;N&4kJgsnB}7>3|8uO#r*ovMaYn78Os(hD?J!ek zWh?n($#qeiq$siwQ3H?_?iX)B`)RxbPa4v$>)lH(Ju>iFUh&3bDbDM6On2h(rqGWf zyP|o}#k(X-U|lxZTrI@HbR@-o!p<~WN^2pV+H%0Km9g2dQv<5X_ZyfFGz&tJ>`VB5 zu9L79*QN42D&{;UvT$R5Xj&!eW~7!>rfT%Ft5i@@&2G+)9L#J+pXgYQucD5qSl?9f`%bltHuzRUa6e zUj0=J=kSkob?AIMJKi~_Q@g_qZ8sk|(P@FI zaI8qGZnooM_4A8gWBo#pHl#pA&aGxj;O26;5>A6C@vKH0f#LFP_)26S#2KfNa5;nj zsb9rTf>xkZ^mz8}1B$v|(HnnpR9`S>@7(~|(W4KlT!XCn_O3YEuP3mueNF$PQo%Hn z4#~26YoR~fszpa}&^f7BUK`oH+Jcg)Z>5aoPowwGZP5(8sDB}uQNo>o6);XeF^@K~ z+<+x}O3sFRoH?sD_~DO7bKbbjY4u$%`39rBX!)%JQz$^#dXG6e(Q;R+m|2Lz&)Vc9 z+3>^17Fmho$k$man`_v!n|g_;8@!7&qf5AQwtlswqyh1p#Z_0?cGVo+685wm&JT)W zcL(kBhHK=rKdb>`wA0W>5Xh-5q z$ahm{_dl+V5H$|`5%zx}TfdR=w6JoTj;r%88hnJ8D#S~-E84JIu>K2~ajJ$O*#1K9 zJ+?!=yHLTHt^~e9m`fNuU?~W(jlU2PkN)epK!L-n?QadevZk@tD@VW=|1Da6XL`X8vW5Wq+u*#;GaQ0ZVanb$gW; zqtbc2kQ;F(KWPnfOF+!_RX;BqFSnz&pxn#f08D<+i{j@R z)p>pv0R^lafFlmJKhbFB91AE8nEck&99r5lgnL*k5s_mp&l`jKbU?cFs0el6Q^X4( zn#I9%JyYm}9e&mu)>yL}C+Usq(?VrwvUtr9XX10ScRZOuPjny%UDCtB3jupPt)!GC zN_aZ|PdEv^#!FV3zzY)ENU;P|LYLMiJP8{-5Y_Uv-t&FG!&S$jX=4NZsg+1`-37+5 z%T5;pm)Ri@qzM2=0?XsO)o7Jn0EFSDPn1CkEVUatW$)LnOkqXZ)832e>isz-Z!spu z<7kRT;d=E#mt{{ZRs5Z`iTA*JsV6`W_v5X z-?OHv{fv5gieS~VEbbmZliLe(fKP6tcTp78tIPe-zJ%g_{&A+)z(s6Px4K|5p-NCp zj`=?neJHjD@Umh2p22ZrtMNI zpbt?5YXEDARpauhDmCA#)(NbOl3WyDshEF<`=rHkJGy`36DMhitVUGa^e*Q?$Pn}_ z$iA-v?ji6z&z`Y~zE`s>nVe``Wnpu&VH8W=hSZLzypNiobGjW*+IR=IAPo1lbj_%$ zS0}LV{kp{bV^~m#hiDlrTY202MZ~b}zdm%84pVle_O^hAFaL`kv|#|#qD2V%3kgiT zOz$@^ZFdmENbAdPJx7$QB?kRy^A=iK?KU9*as9@(!vXQwnpC8aWk9v9k#)FQ>&xj0 z$qPM(!xBqRJ0W+DHYS8{GK@gT6gtNSU?y?rzfxx#%71?CPqt#)<`V!B$bLlOo+GYU zGST@}%20ra91r*UC#1U_oB+Bcy80)C0TI=!CbALns_HB8!(j)u&nT($_p}nH+sDBW zCM^Yvi;(sFc~CL*Q325V7%{}%jq(&v7P4LyRc%(}g81vaTYZAu%fd63OsXGpdX@_6 zk5O-Kvh~t5e$|7@fmn1GWU;h;wr2_%`-3i|tJtY2x%>h%Cmg8%P1Y_Dp0E z4q*9Wbe1rB>XjY;H(l&q57%!N8&+lb0f#qf>#0Hl+IzoyIr9J?=L2#X+KTSkX94)c zejL^AUx&>13(jpG7!dk0;JEzbl}*9lp*OKNwVsYZRD1GEltf2Z^}q{hD$!*}^ci;W zt>Q^!(stj}{{!xu+#j*8z^SDHimxF=!pXM8fB_I!ZS6}~UBj)ElRj7aK8-%ELO!R> zcR3Y5QA$a+7mJ9yyeB4&_>5X_`?Cn^B%23h$NC-og;bUA{Dy!^mVk?1jqL3^xvl+u z6B3$p8zOdWrTZ;WF_lY+HxWz0Sv>8fcn)ym%mK}(&aAzBMk;46%g zE=XPgZ7&DDE0%X)M8c_IMhC4p+GypKILkSH3MbOZO}P1I;l zjSSa=d5{wZy#4q~17`vxV&RsWMnprId@qIocK)|#L|=J__JY;&ZxxCs6BO@6yr+pd z8P(<{CLO}PTjh0{O5uSprT85v#*(QW&lH>`J zAFaGwCSWG&*${@u2v+}H&+ObnTSs8mg1|0=Ga@Ugn1eARKMo=ZQF%pJ1F4qBcnO)q zaYh=rz>_wF@lTIzpSxkrrAnTz-+*fi*@M)Zp2Xa^C{SQH&TVml;@@4Kgan}xc2{np zk<5ddK2Spa&m1P-(?|N5$~H50$_i*~i^vgp9;!`)KTzI&(k`DNt8NxqaDJO^;RlGiRpJF z_`r?_6#ppu2nx?@cBTs5JMkl5M~7mg(BJJPNrTZMamXW7ak9$`cYh3b7xz=LCuvXP zxXVv4H~8eKCV!d`PjHr&s@jx3R^g21`TrQ7%>2&?NgSqgv`|zLw`mny+n$}P;dLQgfB>pS`@S!URz5^PmCfxb` zPk2*~eg5`&axQ0_JpT)wM8OCBD76Mq3%B+CifnD#u~iGt}JTXcR9QiXosxl z4w>rjs#Jee`Tb8v(v1iwkFLWqP523V@d&S4a`3rh&<75;!p3TEKe4IZ+q)5&XC>y+ zW97wBmGf4f;VmFB$$g9_p|?(g@?tx&q&_x?9jwdlUu-e-?kuqk4~+ENazB2MCU>A< ziR~{~e);Je!w9^zuSeOQ9g`@axb#xcqyV*Ls@7rxl8vSfBMXOvJ zyJof>nqzfZ(z&Nwm-ZUwXB;qOFpmx=o*l>?i;OnPXzi$5sq5JOD6B7&E0+lYMo=#x zUjJ+~T)b(?00({!K33oo6`FnSK6wRk(db-QdGpY5itbR#djapWJJ-^5e-}p(J5i(X zBZz;yw?O)$Gg5al@Rvr)kFUbXq~)>b=*U}d%*3rY4Igbs_QhQkx>^<|V2Xb2ORF(4 z0v7NedBqengNIYs26s&9vjXLKhMLqT*TvdJ^O5V2qR?23`nmz~E<%s?67ZB_C5Ev) zwZ25yK+#FF7*4jmmVn|%@?&vVtbB7fbFJise55B?Nl&Y-Mz$h`t3&A~lR;fh>vsHP z*TG!%@J!tw?{#xBJFGLQ63F2%9YV3R3)9%yXpDYZHD-Lw)&Ka5cRg-lO zotuFvvduwo%q`+;M~j8ywsGtdcNm4vOC5W(T#`xa;|YWs>UQ0|@5YmsKEk*&v1 z=t>4T2C-b;l-{gHU4S+~`+0@I4y1s}K-e74vGl9|8d1a>$)TTfW+C#D2iq7Ar|f$n zQeI^|BsT$-4`4cMF{xk{y{F^XeS%zJXH<7`cRpy-_M?qdLsfcuyOHt>NC=BQdKR9A zY7QPYv?Cv1>xBxrkcaIX7mO~UVgg=#(`_n_SxUWqyu*s&ZN}@zoLtvGY5Fc52o{d8 z{`3^HUQgUT5%8Ot_7LL-ew!4`50+dg0v7;OAeV1+>6ugCUj?k3-yp391dKgee=IJM zlFP2N zF09n+uARrWDSd#&WY@y{Ca|U^U+Bu@I@2#K?XmG$pH6xm`vi(-@a((c26_Pw+QggQ zWwwc41fdg|M$QqzEv3PxGxd9W9QU=ybvl~b*%-I-^W#`67~d^bLr^O4=HY*aK5Yw_ zTfC{ryH~|;Z-02DI?5F?%;sm*IMSJ(S8-auS2{%H=Z|zq-m_Yj-=K^o zU=Vl7!|Sk&W84J0IO_G63I{u?g1kFbyVWnxt~j)-?q%|E4;IvpAwS|J6Qa1`?c0X z>D-(<`sgGRY2-Y#xw~yMF!C~nTn?Iw7)t)On}3^bEAjfs(owz+bMg2%p`%7!mvAGv{@4?yoWz$KA9)lfx6jyKV=j zVx{h_J46@5SGhMOoe@Dh1QSiC{ zJ!aao_GBN{K0Kt%zK5uMMKAV~fRiJ|o?LPZ%KdX?KH|d>Qf=s?-P^Nqc8`zG6!rf0 za_fz0aC|i()Nuub~vV3=Z2aA|~FKaYvAlWc-*nRn|ROA2mDVY#_j8sTQ+QJH?GQZ zE$8twzEEuM8Mk{K;0j9#>#Pfgze}{8dHqa#*?oEBO}~KBcIk4Gy|Fc8Lzmd*J~QI% zbXMoO+P&ccgOJ>(dg8GIky@;tFFm#U<%}}6?%A_?%Jo15SX*IzFV?}#fbQEU1;?veU{bVV%ig{Cx;}amG>QKJ{r~U@t z!L>yjH};KAI3CXJq3LFOXIhscU$w(OV$iPv4bL{ht1=9egm4<-WW(VOBcwFJq}>Po zHBUg`bKi#HBWqUX;LOK&4EUcs7(#Lo+6G{p{w5s6`o}!zr~Z@8BDzf(RjJF@4w=K5 z`7XzqF*8D7iDzKS=P5R%HCld4_M+v+E?p||LnooIQT+S4E`)HDHq%)d=#Ob(L%YM| zd8rT+wf_r404M^3Mv)aT)JHr0=))Ei?evQC4}SOphAOZ|zi=0~K*3_@C3`QJehOor zlrLdOJhT_xvh-QxNpd3Luy=gJm}$T^;BIheK=|!qtzbzt$uy(fHz893Asp)oy#Dlp zSHJ&du%ytZ0HFCZ8@oB|DNZ;XP>Vb_Z3I)Rt1bnc1UnyQhvKW%Fon*Fo9eqZnyPmg z^dS)5sR^=z#%=?#i4+r@`ec8Q#!Ba(LT4IaKdxHi3L_~e9jx-39?PbeUnD(a7a#a>((q&k7X%`o za|Cr^ao-$bZ7xIv2*}RnnlM6}mS-YMp6M0Ocyh7D$1hM)#C~%;)x*;)SWZS7*VQf& zz;pQrn=~H-=?GE+)q5G{4aibEw&4oXhGFA}&QR(Lbx z==I|aSLB7H<8*h31{0(jT6Ng>w;H?P)T)+`FEn?aM|)qW9XTRjP>^hji)Vc+iva|| zb)*NW5d4aGYaZ{{KTsLL-*!mt|G zeHwEEhPA`N@QT-oDjTT#Dap;o!b872Z5|XyMptzvRc6XrdFSk?8C9tz5C5GYv= zHDr}Rr2c$3p4W9|KWVVOI!5hh3?Ap_6(-p9`}XU1URW!6$pmwhE&tF9{r!bg4CE7D z6mI0?7b(2pYf51Urg$ttk&OLF)dI?H|XG&}CPZaF%RrmdlKffNd9I>zi%)Qr( zp{+M-H!`(BFN@piFz-6#S1pQkj!0k=3{luxJPAz8OJ9!1g|PRaSN6%ufDbQ)(4;;5 z#qEnDHJK#7So1HuqV0}t$O;YCd*tDjgP6*@jCg;RA%|@orQ6L*dyVoN7a4RlLHiM% z2Y!tEg>zO#inpxHswG2J`cGv`rh^P2;LHK~-Alla6w{!HhY}-6si-s*CqfO}aoq)1MSuQ7A)#r{f2p0>xzs^MZ^RofXRDT7L5B+j-MN~Sby|r3PLz{< z1u$6rKb*aJIFxT3H#+t``%c6N*+Q0(A=#3~k}Yavm+T3Vu|(M!p|U1xVzRFpyHJre z``8P^jEc#O`JKD>T-W=)=X%dS=ls!i^~acJp69uj@BR6H7UokO;xNgS90qt5+4#@_ zTZ-zGZ_@#X5C3iVqqnqMazAkopdBW>`;yd7YmB3szWYXZ@}$Mj0zYps4zD16YWUaD z(P{_W#4~r;O()bkSV+|!9o2{nG7q**S0rTb44-4JWDe>-!=Sc@Xi49bBi$q$omP@v z2`sDIae4hh49{QBY|XFD?<_*N=4I}PdZ5m;3aOYtQp*D1;*rTD)u^eWMkg5H5m!4< zq{5zAG*30*)Cad%a5vYEm}=09t_f~oes@jMHe6KfE(F}Fh8VXXd#-_$m$VZxod0?e z3(-@;f*GY3s;h;c?dfg3uvqvea{F#X!qroXy2t*fsnuK%_o9FI<(M7?+ ze1_&l?g~vso%_LXZQ@NwTS~)XxM~bEf#$+{hgU|yg~FTNv=o2P^#YwQ-xN;@ukZKu ziupa{U&;R<(+K~l?v+N&koJF&7-VBw6`D9^Idc#xyjn5+JF!VHb#G?2$g^DvWtOh| zT-Qo8GT_gf>k?$t7D+WG?tB9x@|a3btX`~HY0rRkl{k9j ze_BMA#MUq?Ab~%3*;x7gnVc5oYKFG7@706u%GD(Tg7yhtO}ZtaVeES_5<}EfA^mAZ zB=`D%5HX@MS-UtaXw7V)y1F_qsLb|!iiEt1$>)K5XJ=MwQ;1zKDIjVBA^K1Fhn%ah z{YLhz^tM}8olp#6b2)&&oLgL4oFDR%e3ZkMLhlhYQY!1hB*>Cp0N`P0QIl0>8;;n7e5>$Ox5J=7aur`UE@9TNSh zUZpDIeIOShYeU;xXYUA zFQK$)4K2)6mAfHlOut@aczpT~qigg&DXZu&isxQ(aB>8wGUM3-`kyXt)|ado1>e#| z3$L^0&A!`riUHEL2S8fZ#S_&-5hD%oOFC6E0E}GukG*IO8c1be} z)7e^x7w2PSg3`q5UL&UC*@5R~Ia!au7QZrd$KNzZ_-EhzZF~3=U#jI`5e->)nQWd- zo477L+B%&gb?}=^y=NyW-`m}nJ~n3c-7DP*f=fuM^rP8x%0AJ`$oJ@L(q?F3?Y1?N z_xu|%ygphU6+TkgOEsJQRy`^Ln~4)ftZAvS?@OKIKK<~bIeA6&CV|{P4B#uc?25&S zv39~kYl-J~EwoCt7P)XADU&zPnz}r+F=C8tV9iwIiM4wq)Va`$#2fiX-xVboWusS! zE&W0@U3r0`eYZ_!zCkILi=2g;aqm#{xKa@}Y9`uHB;F*#j0BsbpGMb|wV0*)TjkJ5 z@No5C)6?Qk9kEx5RZ(HwVT&r2^s|5i@ce7~Jje#A!>1n=5tE{%cOUQ9te(28Ez3n< z>Mp@~z_7H))E&-D`k==X4{DCVpqoW|fb37Z3s~%`cZguznoX<0xSWU$=1(NBD_TZX zuCnxA-<+8~KUK6x*?BSBL5s#8#F*t=k~!g6o!MM&$FPr_g4qK*&FU1z=dLRyqa)9)!s+kGO#G2qs^~)P2TP7UmrSx& zdhM3^%2tBIIrW_~qUsedu>lO;cTFGjuq`pQzPPfSWwuA~%hLiwXGxy76ua*Q1@TMg zJA>e6Oatg~5YNv?KE%5p$LXr>bV(3I3*WE}=2jr)&F|43;|q>PJR>s?!l;Q4XL8!!0Xn0FB_xvH;i$7}Rvnj#@HX z2EG_MhTDFdHB7pkBqqDWwxGo@xHKZjB>0sP>dH$xX=6#y8D|i~lX@O&^rh1P!V6jR z{cF2~y_(-7!$*p1EMEPA39h3Y#cszsv-c)mM7FKvN&8n8v!_Iz16!}&e-Jlh{PM*q zkR+jWcF6brkHc^cOMUn7N0}oKDc0ukVL^A8oXmr2>i1N%cBMq?qj#r8I^Ot(Tg2Jb z6(NUF>=H-&3gS96Zm;$wl~jU_sZCzcUoUGFKZLycYS9M#sJ^~y5>U$6lhx-_J98$A zczsl!W-}|NISxZpyPeO>re?nYDX}7)h+s1HHy!;zFk~A}i?7O*FRc_3>-eapq_bNF6LW45JLgYx*L^I!DcRac$Co}^X*L$@;3^Ep% z`LXLO;O?#y%P1Z(?GItZLy(m&d_D%RNC72KMdNj}#y9TOAgMC4`i zURI1~a6zE!nv#bdYq{MBqEu}GLAYN%7b&V%aQ@cEmmdpKQ;v^jzB=V3X}_q?b~n_K zV~EnK+eJ)jKbfQ;&-@_^gpG_^PU_sL`FT>)m=oIE(!y!ZTIq_CwE}Bdipu*^j$7Ec zK^Ynn$LTZRjVjqw)Ucmj2E2C8`SXU|c6_z{LoRb_b>kB~3PwqT+y z7~0b!gEnuhp8ef9$|vP7q5e>&Q1pG3?yqIn>UR;{xg~ z7!zR0D{)17H3GgTB1r>a$njsEng)7T=t~h%jI9Z#R+`d>V&2`fwVRN5Yy;>T;LR}! zh`MJ_I~~`)L>vVX6C>6!TumvDX0K8%E1X>(*syaI?db9Uc-t`LQXazFqPWtvEJ%dehg=o?(-UE?n~ID#sEDtGBYn6n5k^dS(hNPKh!{;?G+EjX6-J0_0T zeedBrGj)1|M@*%1+XbKc-Qq^^)`1uHd*J^EY(sgA!DbZEy^mxg>6jC4K6Fq+ ziY<1=>e0K_*FQ>-In2~RUJxs14e`aobHvWCRfM~kS3^=cMFWbw;|{ff$MYpnh0 zR;IAqCm{A*2~2meLbG&TjD!>EDjdl2EgjPejbF?a)yx$K1h?Cx-&}J}{2Y5;JBf8$ z9E)4uL=r)b*n>`6rX#>oA|2M*R4wo;PyP2wS{(0OWfIrO)$cQwxBNs!4P|9nX9%+# zUq3F%w7KPL0TG4#KCz%}`l}39VSf-_;j>oMQ^Z=7aQn`3htHdx^x!J{?*}=a#>t~e zmVVK%uDN~C9KJJ0BPjXj<->RUv(2~N^N&xm`yVwof0{6pujSJGMsKF7&Z{Sa3ShVm zA)A4bqZ4@+1OOa~iH7A4JvQHuSuMDy;L_0ak)6gonA7DK^bFNJXYAjXkgS+|k*HNx zDjXuW>-=e3)Hx6icjdSb^WDD(hR+*w>y<S|XhrahEV}rJK5H#|1879q zKnodqmiAutJiEHu3n)V2v(v_E2^1$FmB~_>km{J>)C#Vz{--#0^urBBt8 zGzS-PV-|X_mqK#CCyzkqBi4yB4%RkYE!kl8ll4%cNHe`r>!eNBG>>!sE8m-ln3R{J zX{K0epI^Er%hIwE*^-sO?NyY=5~LgapA--DhQ$AY{a7&kP>6Di0$0JC6~%$G&cGGN zwXU)?QBVD;rXH5Nh92X?EMJ%z`=iZ;>rqL?mD)inkgsyTmy;kw|09>vj4t-5zIKLW zvSAr9b|Q7=xl3awD)FTLvNl3!)k;lEla81aWa16E*_ThI>r{|iU{4*NA ztalktF0Qx_?J#u7EdhmDp4_R~(H&xgYlY)*NU%nNnoW`n$FO8wQwHQ+F$(T1M|^pK zh6C`FXT`E1NM8&$@_zapm;?uLzd_iEKB1tnsDzlsV4ibMYT-_m#pY_#oLlBw8$ceP zFRgT8k{~QKuIXcAb7)SXUr)}hh?lNL4}y5OOlV_z zhF1M?+i!h<6^iFx&hfl2aCQuPye}LFRg9IrE6goqVnXGR?T~$wiP=PezU!YS#KSqn zT?+U`tgm~Q9b@SB>diMLVYyuMBcdz**8~NnWAwx(Li-U5zmCj^*6VD>-MW_u8;{zT zQ52gg4u`u-?^XUNH|F3?&u;zz#|M}!gm5b|lwfdV3z}x`2g1vUu&=VZ!oxI;YY3!%YiQPBJal4T+-MPdul&gsTeT}JT-0f%0_g9s zc_m;miiaJQG~xX!bgPji>8#!sKXdqnoPNq%dsk}cJ-sQ#$}g3p7u56IWknXT5yksz z)85k5^W4>FQz#y6o$w~)Lj?Wa{sUbOqEPPvVceDXDqc}G9 zq8H6t`dkEssR^HVXH((7REpaTe9hk) zC(SDElt@s2{`~Et4J61^hiVxFPknL4!pU6qVjMe<0?CoHHEXduUNf2x|JJ(L*on&r z*CflogYF@t_EcN!Uzb9qcf3=m7(dQ@PW}Jq02~C2P$1a8wWPn zyX}3%phFg-0%*3BgyZr}nKIv|DBFbC_Gaco0#w*NNruQC#VB^l_zF)GEa{j6ohHMo zcFh15#V;qo2q|rDshbjhDX(&PVL(H&@K9*s;y>{&r=swVV4`mY{&Enbydj0QrMdOs}ZEbN2gn-yrFi^oN<7d^a*1*k7nJE>=RhcGBq0>UwYCh*8Xq1G1~_F^@~WlQF1 z{Nws^6OjzMCzgm|5G4}SOF1|i!!J#jHD(#OR_xS&{gE=vC2T5bn-P6AcJCb%whjW! z9d>p9A0$fHf7AaH3ZnTXNnsWv1p@Z5KAhgWkP1&JsJ?$zMk8ogz1-bPB6f1rDw#*+ z+Q3|yb!!)otgz_=G(e;4rOCY7YsW*#bP#WCy1v5n!}cUJF_OsjVthc&-iqu22)!#JS*%GAuh@Vu2JlyyKW^OV_MPhrqf#p`MkP#EWaq#6 zx@*ENN@}Vd!uH@`GAsc$09VvEK&UDsTo8qsev4G`^D7$ZFni^ZSsZ4?YjVQnw&>7h z@h&2Rk)8p^0IE4d!b(DyTG%YmQUecJ;c79qom?%a;GG%kyLUz&l=?i%+;CVGc$AWD zQT#5W6s3x=i#!iT&9mLtV2dHJtQXc`?_6NB>z?&4uPwUQC91gjhe^U@v43EW*SuS$ z{1?W%PxgTl0FeO8V)Cen00$tuFtMSB@8l7>BrMskb6R0v2&YiwzE=~t*?2RE)#K{Z zZ3aJM-anzeNDyZ>@E?Rdsu#f>MN3RL8k`)|)$Dp^*&KZt#mM&bgQQ=c4=OMapf=65lL>9Cw+MjOd0x93~I;7!AO zjK=*UMTMnfJ=(DTEJVBRlk;tgm^VFWv-2G+OG8|DI|efbMBsd)H?DP{4Ctx!dZd39 z_EDHbNzTjn`S8o$^XBlOlp(r(ps;oPX!=#|n!M z&q-Lv3xm_Q6=T8Sn^p7W;ixq=)D5cErfgZ{1okQhj|QGYXYg&4d4Cv z-vLY|o>GX>tSzfsv9~tjx}^c!w1oKXdX&&{S{?=Y9;rXosEuI1JHIqm4)<}*^zQn8 zj-6Wl`Z+s+tM;*spi@vc1Ud5(orrXX&259X+yL7<_QsQjayiUu?;sX;G&mPT3EiHI zB}%Hq?-!^|fG?IC+YqiDLUuZB`@g$a+(0dWPMwUDB8IM}o;?!P`nnvETx02X-rKEB zU!3zjw;rdZ34?aBnBL{z?8JXzp6qMvIG^<*?lAD5XTU1xiL|*;!N;$;+eYzxUaMXk zY2ohBEE#v_ZV00!loN{p+1y0w+@XVXSR*ZnyQA=yB~+B(KT`iN%yTT~0}?rT6Fg6B-Y=?}lYr^7-)9rYgKJ`2wC2j~~Vr+2Sehl-Qa+$5LIgy7ojs z2Ai|e#PBpqi^x70rPeh0uz_Dq!SK(+ZWxbK`6G*Jqvq1b?S2aq)>o&G2hef4YDD+)J5s4DiH*slj1}2tM(H|+ zJRbzKgN#fz&J4Ywc)?aD5Y6_Qlz|D8v#;}EQ~yE!RMPyRIr>)|pu&c*+r8{@T`)X< z{KLO$N=c9JO&6$1zYdcxF)4oZl<@|eq+kBMV+eI-;|AvjSlNxqys$N3`I%lQ_-|!r z@e1(nlJc&Zd7_}`T*!nvpDptD*XQ@o?%7)K95VVbUg5{Zfra`H@e4r-o8EmuOV;@~ z8RCmrje0R}cxWcuXci~0zPo+(kXpUkF|+A|iL)LyW?ih#lgw=gRE0JqBck=1fB*Bw zCCO{}NqES)B%I<#!TkC7d)Jcpv$p&eLEH5YpbsB|N~TAbd-~w!o9Z7q8ZBoa-qOuU zcH*oMGg(cVbV^fLF3x~pI|N*g^CKMy(gyg`rr(cwIH5w1zOhxCMrb*rU%rDJ_L;R! zIe+*hAg%Tr7YPs)vH+khiOpdKkxY@wA#CL}H3ko6(1DrUX5PWP0)5LD4U`R?P0|d4dNsr&Yi(llMJoi*G$(LuM z_`q7Ml%S*GyGp98mE!YlA1;^)QOcH6JXGJ{6%Sz$r$HWpLN+#seDkzSSHlpb1SrM- zj*tlv*3L`zsgZHI`pxuymn#!rxaXPL;+V!N>zy+UyR~nzVl?-U(<|?HiB0bBZ;~cP zeYiVZ%k8VA#cq>tfbaV^(XT}5M$()XixX=I`tp}fHLllnB*Jmc653Iq8ptG_Mdk;8d%sFAA&cZ1zc+8qrcQTe}`cf$NQjbk~RWt1q08= z1gJa4fLSK)%Nl$hJ_Jl10@q-)Kkf>t3wcf7YaPFbnmNz;MTA}Co5>-c>$a>dD}-5} zK|Eao+{#nXq|wzOhZ9dV29WVS_bPwR*Tlb98MrcXd(>B;G~kY|{s+n0AZGpt3HaXyH!Ku*)|&BYFY@=ax1JbP%WdDw?USGn17Pr2YqEVk zSQuq~P=q706l<{ITNC@Pp{%_UgN`l`LHp0Y4*lxwC|A5;bK;C2CYm;O*3( zVW>ZSSfbUWLn2ksb7AGh5}H^%z;RLWFIY1Xv(FO9#&5#<^d`yh)nsDG73o{?iVA+d zi>5qvCMvZln&NSK!mXE_Cx@=fP^;+eb^}j{5sV1s+^((+_)*Uz!=}0z!<4MszEKz> zHCu-gm5R^ZY>7&<+Z+GXMt>2QwAYTiT`*0tn+FNX8Uv{Y*cM>1ff4qqe;K0hCI5cO zNwV~~_5(CILtptoDQ(rT@jT&h3JC_uKRH`ClLXxrXAqmx7~n zwF-JvDG2B;hD7de-7`e*)d+TaBDZOr(YN2x>W!h_pxp$7K(x#0^w!c-Q`gR} z=aP9lu;`)mq-9?GqYC^;?=J+N5_f3}G@d4>%ZkSGFAM}Z>$mGA30{16-B0?B3j}uW z-s}0JC6DHZg@Q&l83{&b`vZdH2teo$dfv3Z#v?;lYwXt1J-%uW0WgI-lEzNrtf^Jw#!BMf^4U@=9fxwzdxq?OBn{?UB`D z*}}5Wehr2|LGemIJ!s4cr28W9|HSM6FoKN%@he?s(shz6h^xT6^kMyRW@`s5KRyWi z3R`EY`7j8P*(tgcl7efwN_ACD8^5#}bQN zC?GB9spIvIDavQDkzZZI+-40?!hVm|H8p=zcU##%ubelD58hKI?jg z8hI4Vs6*X(G#au(+vF6a~bkP73e9;FimJO4`>iZ;Q6uJX`L|a6KSiD#44AQ zx3_oip=;!4<~0S!jm1!IF0(9kql%ErPK7$Kqf1~8ya?M}%f+#;q7w?w6C+mdD311T zIqC<$nHzNvVU1xO87O`2OrRXFzO?@2yrRfnPMN{r;fNF%R}^*R89aY9`n{UD&uUbG zVs5wy>d9+a$~V!jpCzapY?U)aW<7av>s`m~pNV zI871(Iq$RPo4dCF^sksTqi4mAh@n-|Ee`K)#+ritZ4*d9an#2f zQJBXFDq|LaluNY3p5sw@`=Mph&-vv$ks?KdpjF}Y?4;(&uk6=PnNAzQwpSXZeOk0Z zjBv%z_O?GbB=+ew6KL!2Z+W9+YDXEd*P3&N{9e4HA|8X8Vo< z;Y#20oAQ($#=CA+KfCO-88{f+P2aMo2G(a~ViCzC5(KCi0OvuWEBUv}3O>)w-o6ef z?q)uMFB18}w^-l5)0@8)msWzgZ+xB1dkPr5g$S`O8p`+#l9wV~mXKHWqd7NlNJN7v zMCx;kMdP_!MgcJfd{XJokjtM=e_8`xfJD>+tTCYu$T?U(5(`#w5^O#;5u-jV_I?GZ zC8@T>YpAE^=5ERUwRwTkTckH3d*VT&eDZof<@}^1fj^e4rkJbHZ@F0ZKJ>#^7Z0kZ zL06`p*QDvo7=s8cfR_7X=6(%|HKO$;f~zt64H^aJuYZIfl}|8a-oCZFZ>pD9=E-t+&7_@hXgW@vpL;R>5i~rD zY21VeQ>C*QlJ@$_aFD+Dj8G1rXE_STBL_@>Kfh#_Y9;vQ_1EuNwAK!Askbj<#N;1* z<1-w-P%7&osYX875yP7t%%89s9Zte08><(5!tY45-ZA2F$h40Ss7Vm+W~CB@U{eT) z@rpfHymddUVS8^5yApp^u08gb-GpRY&SuTwH)w8)%96lz`tl)pDenV~tGM$x!&C^- zm2?l;V;!Xs$v`|^gM%8Foh)7Ac>o5OErxdIu^z>*LI}?To(mIIJ#Lm z^dKZpkY7p)4qG|;sC#ijCt~#!W-=9du{mo<{na*uX*PNr{ZO0oeP$~4wE!)h1Vs6Fpk5N`|TDx5- zNR%6>=S(m@Uwn=bh+Wk776rcaDb#}NOk_rx$aao$lfxMY6RJCr5d!%(Ww+h%8fwZYM{1yF&gX9Bh_YfIC<5u& z+-AUY00E8+#%sp?W>2(@zeTEE%0=mf1}>S&G0v;SOFqx%Q!4m;HJ%|hfRW+%6@gy% zVssJazJH1enVLuk=oz*rw;nIHv%M3*Nol)e?|Y;z(LN2{Ovq48E7OO>>Z6ONuSUNu zUhDEZt?gp>!JGy7DNLAr9zXC|{c&@L4@VPfT|!c=so(BGiqZ8%R!0qI+)@XFfVkLL zVIFu4Zp1h4Pkg~n%}j=HFbE_tGHBn2huX-cutYOP@a$O&u26%k7k0KK^jQj4xlY=Uh{~iz2_Ymv7r( z*iStnA-jZ5?c&DtH{Q1G!PghD{uz9niYeY4Y zjl+*446pot$~-Gdt!z4FYx$u=G?y{zeVNER6!QkJfFx8*=3ncVcJ=f%g(}jZj#z}< zHNSq~G%;b==zDb!&cl9T&)bsk65SD>#Z5g{lYe395$CCu9#E01SI3C5{Lh;Xm?;Ao zE;R((%9rvLfksl`m5!HXoK8D@Pl>sS+;w?4XlT7h&c5+NQ89NVdo zwd$?C-!F`D8$6-ig?h~KY0lNjqh7wIet)aZDZ>ClMb7;mR^9*l4z3HJF)g;akLsay zW~_&jZ{M)iz860CB3fT0mP#b))B&sIf$y{aw z%Q7veHG2S4fWr?oP@k^g>Y~KCney1Q-w)W^E^g}?7xl?BIXXXf5o7E z>_X?`af=ghkajyP*?sDG(=qKKOvs{-F!=8b3 zQ)6!W{7T0nhY%pI`SEGGyzY6jMrT_+F~UD+X7#-F`KrKoBP%*8MkZ_l-PGZX9Q3qC zR3hN#!mu^4dSqfJyduo2tEEiT=wXXv^@7aAymUsqwWQ&tly^7juV$!bMKeLfD%Ok3 z&)!@V5*jN)M?r1wLv4(Bj zHlF_|6UotojTrr?yWV@0Ksi#dqH7Lx>(tq?+^{wIN)yERX&samKC|lxGGAQ3uEe)4 zx-i(yZ%qKSyJF|0yllW4&$lWWvyo!LApZ`_!7kXdo3wub!+t>2->LbX6F9WS@*>GL z#a#61RSjm7GadW9M}@`N3?jU=GcC9{0_;ehI6#so+O7EyI_BXFdQMK~&EE`{%ck4# zyG91f-;dIBWs&#{9f@oe^i7C{6^YHs4K6ExgWXq(Z@F#T67Hign|P_(!9hB}Ep~rP z(<8NQ(=Bz5@rwS?B+23{gBtlq)MtQ!_9(a$N!UTW0w!64pg$#{1wT5@d2r@!(Qk`h zF8Hf0oU&ih=7peX{xdS7dcnZ~l`OQ~Alq>`AVg!^wCj`BGv z&S>jf2msOuBr00u;iV?RHt^9+K0*?=u`oXJHT+;jV1;)Vee&4!`jqpbM|1P{3Hb$A zZyHHjXLZz^6(fV--{;^$u1L4>Rl15q#w(N_dYD-j1sM5$6~jBi`NE1!|AWL^uj1rv z+L|W)jnG#H8m&|eU2fsatI9uB!dUMoqDSUPAH^*p% zLMh?x$bXPd=42{1&H&<*g#Z-+_B@P>AQJm}f-?S^lX%5{&5P_O(W+%ipl53-+RUKX zp#rzV9gqGNDK35g+dR&2@cbv@47f969m85cYf~a-)r9$2kM6HGom(U<51F{t2(t8{di2wy)9Uj6#vq^qKm_;_`1e{*tgJ!jGB+M`jS#w;V# z)E=XAfvqB3^gaAulx>W{xPeKunG;lVE0a_gE0y=``h z7jxU4%^-bR2>ZrJ-$2xb34h!lHf{Tk#*XGpXp+}<=3Nad=K2hMD@FlHfdOQY81unF z6FlK8$5hkh`K_a0#*^uPNRS9mj~!>$`knUQLAmASgwNUu?tA z9bYUn{d%gj74FR1+|vPcEWD&}!Vpj%y<>c(S=i^!^6gJ2%5)RKk7_2je8i6{qsH-r>*sMM0F>d}m#AHhXPmb1L{(Q>B&6%T6l0u2MeFx+8x=VE{ia!t{0<<|0 zpnajuB4|ACG$ctmf_loJa`C<~eC7bBAhc$bG4LRfrqsj{D$!5}2~dqzPC7L}gSZ#G z#_G!6A1n(zVTJX<-BFXq&=-H}1>IkVnEuoy}8E$N%i{VC4%Su^MR~qAId@FIXH?t^6LiSf>`E17^1{>vW`tKVREogxOKX= z{M#0OF&FN2*w8!6argF+NSWF($4vjj&*pzh5-}Tpb)m>buv41b zf*!nZNcm)54KSTK1->}3mQj-{ygEvRIK$I|5+n{uN6ay+N^Klcs4xY#RE;yAfiFI!kx_&c~2zM1}*UM#S= z5`v0kfXBNDS>(!So$4DPK6(@Zz*<^)h5E z=%Sr?T(d%PsI(xOmKaWoh?*&4b!}NuhLc_OZ#ABr4H&SQ+aM6-2)N6$^FN%;>dXJA2{ zfkS7f$th%8a5o9y5eRjO-AA(OO8Df2o}Bq0!OvM<9!12 zfVVQv>dXrZ8psjW%6Td~v&Kiu?%BiSrHH#7X+!3dZ%}JnYaUD@7X2*R4QCtt9-HX7`y()kT#PP-`6#8r@Z*R^h7?`dtL56FH+M(soeJe~ zFAV%M;IxBp@&tqvv7WrYo&fvh2=Gnd@UbzZD~kMB{U7A#*Bwko(*K*ske*m^6FC9d zV@md1s~%fWU1N(sst$n)-I&Q~&TRb3#i%>X6C)|uTf6Pijm?p+QacTw1!$=V2AeO| zfB=K~xm^)=r@LpBgU@S}QHy`Y(sRjzHDOz0qtwCEm|L0AS5s0pGT5VG*u-Ty!uhHY zf^}beT2*jx*}P$yQ;)CwdxGI5TmVW-7E14)O2e5FPgBfG+D0f-zO2+M0bk3zqjrr% zJNKZhU9|AsJ9oeODMC&3IfpWMZFq8=snhH?3V|t71rHSw)Pru75|Kz>Gt!1NRE1F3Xz56`TUxY?TOr(xPfiF>h$Q~{7y|oVbg7E5>@`CCmM}x{c zqrSmH*L$Q}yxcsJ^ihIL7T+&ZMN4Kdo&s(_EUIV@G*hsvB7@U;>k+=R2~1E4+k0j(GWSjRJik8TDc`b8`e2t6aQRVK>G;std0UnDAH^uQX8qBhZHRyV~@ z@0lprst(rBWL^+#Ks5fhQw=sLqN{caMO%H$hA)vcH(Jb|e}z8RK@FN2b-#6Xz9=cp zTU`}3iLCd>8IWWF`E9UJrf`)a`jhanyv;w9DdA9%XAn<&OI~Ltqu<%xMFS^D%K?F* z7T~;LMevWw){}o6I~=P+9VA{0ZWyPI$G%Sy0QyCJS`NK|C=jP-5gD=Wvt*2|EvnpY z$#Pr_6T30Bnf}$?xU1Eh&nGrXPes#0#aMsHN{-(!kHRDpA}1N>zy1qSsHA~AYd1Ht zzhNSQRQ1q3>UI zN80_{AAlpT)S9_+UHD9K)Q-U0C5Ky%)-qI=c3AFU{R68mcy{LiJFK=#t z)H(dhUTTEWvktT+|ojC1Z#hgUgrjE6I&D zo_!~{{Q~5@P}&b)Soc$fOOs<6?@eM{$ANYBs!u5C3bMxlbI)7q+|aeomyH;}FN zO(3l@YH>~T?JdhWJs?1-T}Jju@*}3cbe`2w7H(?>6wrB|UuRgU?S5U;9PANaX%Q6s z5;YKJaFltP1(*rQIk6tyEBNf`e%Kkm==yIsY|7}9D}9}*0chd+7?n>AIve%1*zeG6 zTAJSyN#q407%d5}!^fqyvDA9G{J$+wC7v*wMaZ(G7|X$|6_nUN&sx_L&bXO}^x z*j0&6qcBD-0=mV)P&xwwf}Gl)=dyDa0roerOi2d!}6DClxPd}V?kiAKqVWd^1z7^5Phi$+cP})f;WVP-0xfTnQ;M>#+=lX2R zGprIe0cUXeZX%IQBpxsY5fmZrzF~a%O8?Jwy@moPF=gmTKZs-7Gs8Z-_XxR6((yHa9A7@P|1BBToH;VgKJV84TH{5fIG;>Xaef9p zy-M)00*1`Te`K7qBNy&MVC%fOQzQ42@Iudo?s?j3-u;xUCev;#cXT+gsjF=xG5GN7 zQqLg0WJ4lwE|l)YpFzY2aX}|e6+t?KMx5cS{Vxw$xI6rXD`|egDNL|+c0PU5J+H?X zeE7{O?w>@x6~$rl2j$`u=1nDj@#J%*6<6)B)5Q8<#}X-*3BU|Zf>`7*8!|e&OZsB8 zS)Svt;rZy7Kv3{FRkRjW^gm7RGMML)cfdxkU9=vKh~Xe6z(kW-(m%8$s&j7o{s%E3 zDWPt-z?HOH*cLQ=yBFnt!g-NB(nvQnKe6Z*!F)OmU145*^O(zVZe_1IbwjFX#5Ogl zO8Q)I%sAvCgh2`v-j4{zi8|&axt$NhXI4d)=KritZCyfD!uc?+4KJ_8w=qIJ%UMYrnE`tj6^GcjWRw>#9&K{8q){R^9H>uth!Q5_C7+=iRND#lX5HL>Xc_Z{P}j3uf$w#GS{{sKuq(D6^D zip;~p)v}_`(H#^JkxtA+f@jiBT=Dc>&HHWeZBqRE39b@UR z<+O>QcvA9FGI&96C{sY7u)U6E3RAL}{Ov2&*FLbH>+63c;AdjS^m-*6W%2p;)%ZLR zu|f|)PpuHoAK?Y@)vFH4JwtjfkdayqCs=6A)4rY?Fw**OpeD!?9lT~Xo=?70|A zh~+A2s6M&pkx($4hu&mnf5v;>%T!O#jZ1bgFhSq|F?qdWxamWXGA_NMj9zy=j>vfe zj!8hbOPh^Hq=3XvT+_;)Sx@=T*XG2IHZRe8Psh%F>g#|McLcnx>JfM5f@yb#04t(p zQ9@x`vCmy$Og*OVW1(!ZkGHCDb%wD)<_~@GG<`{WOgmgY|7dI+c@91o6AjC~521K$ z5giCpNNiS49kN%oQlXoG+ND=&XnpZ>kcP31v#4lPUT=|KIO+kfaP=>Z)!9De^Rp6! z{g{#;%c4{MN@%5;^#mDQsXT5=wp%IyvSa%9mKzFk9@CaulEv^x4A}a}Bk?u|)M3l8 z`Mz7Tm=9>gjhTnI%kTGX`B1X`7v-(uLulQhR_`k^d%Ed@0Yu#+-z5~bL%Ks8ttyXs ztWbeBdQ&zackSET<)uzIE3>_#+dW|ps&$cd2~5TLtATMTk0HpHweTUKNg})lC`^-V zrP`kG-|YBkpU5b(aBYI_lXdjJCh%??NZU^EIiMj<44y;?udwfMICbx#45dmdfBXm- z838Av9)>q-nA|Px5X@+-#J>_n!P>oel>G$J?2uYpoCSBq}6JEya1#Qx{iXYod^OSO$ie7Y@u5cWDlI*%ZI=tt%eHi(@gv6RPcxb?~Nn^f2ZQ#c!5X}Uz=2T=oybp8s zJ~-MQ^tL`cIJ5F~rVBcOK2JSO5dBy-6>KiY=X7C6NR3r%f(gP!XMqH%qt$?783;E@PgCr>;w8I=)WeD zoXPg5m3um&Wswf*ExBEM5ll4)3MG7>3CoEwiR=0w_Aa{HvljH9jm6krRn!V5dRBl0 zKh>g^<6dxbItktTMpCFbNH|zWvJ->MVHRSZK{L7Obb0dJzJjTr67e--Bi-yT)KhIo0hlWE=8h%(L~?~3V$<-=%ZFjg3B73PetxFhi` z_LSTu)*0dGPvbrBh8keX<~sSd;f+-g&~)QE*Mp5a{H=DX>pZ)5Oy=o3+9Ixy)X3LC z05G6b1b_ee?e0(5e5f5i-*AY0I`r^rT|77rT5?{dk++Xa?F8VLtuduPn(P$2H>^?rZUh`Q+8(5!WsMeRzY80YY=~b`1U)70yE-WfgXEzp6y2D5un?)iH}QXQ_nuu%h26Sv zsG&&{kRGZiRS*OL37|Alno^_%rGr$Z2@*i*9R#F2-=55>o-gnVn3goUvY`)nc_bcE;46x0N&b>2|*T&I>igCr<3fTUZ!0W-yoaJ*W4iqAF4{a+)B_HXEry zJX*w3s_2`s{dJB;)hl+NIysE}B$PK<(_aFFkI{LfqZe~0<}Uwoy(JJM6u zyQ?=!)7#l%;q(!Hcm<=jG{Wx>XcbBo%V|!rEIS+iYu0ET^z{_XleFh*jbFi!+Xsf^ z``;zq5$v9C_A0Ko^};OjzYqJKqd8b3%K1I%c4$}v{2uzmiZ5&53m$k8ep|Ea&QvOW zNw$jReIKyysyO3DC)69uk>{z`yHI=-(ZC%yeY{-r9MjI?Ul|MSAq5`#t zj5HTp9)XE>2(}MmR)!0#_l@xuGJxCQ#c^d3G<_qjG^l)U(DW@F)kO`Gfz}2esS5Q^}TZ{z7hp7$3d+5>8sb^S&N3PbXn8f>Tg@ z;ayCr0AlrO^=A+Dr@yV`oAvnZt~;zu%v(|caT!Eoai^5TEskv*(70lN>8OCmA8L)d zVp27XOI)6g3WgBsMRTDKp{y_T)TE$rd4KP*Jy2xoVI9}bm`%#JR(%;fU8`(b)mtk< z8&NsL9p!Z4dX6SeXYer??Zd$3)266K@Rnjrig7I(LR`ew{@@bPu3QhEU%ywfXgLz~ z2Mr0Og`RFBCuMlinU>B>IQJVqARaJKywbFNH9FoVp%E$q&=?DG-`08Jf%^UUY%8{2f8u!3{H;-WxTmV9k|N)S3?0E5D6LbB{;9}$WS`H?7kqWQ=^a+7dq1y=14MXS!X>638G<3tN3?!FD4U z<9v)f{tLMdA$5W4WEgr2_!N_$Ekzi;wD(M{+3j4+t$o0eFy=e8JS1r;B2={j4oGz7 zkj$@hRMC{k9K6B)pA+rSS%NpYL_td=LM!ddZYO{2%VOG2qp!nIbO_HggXw%-_VOMx zLkac^NM@20s4X7k#=Cf}{Ikt9-yaQ?baI3de5(5xO2jrEM!I;V#K!4kXB$`*wFTC2 zR+ChrvlRZ4-2rnC)yAqFv}IJ=yEOVNz7z%vvxc&i3U)qI@h83F8ZeyET70BN_BwoE z+_Q~&yUwn!6YcdynND0aGs~2B4}~c}b|Qw^&{x+>6SV{*>mE}9*g1>13U_@M}f8M%-s(sg-yz>!X3A6E%pDo&x>(s}V zx@M3^duaZ}`sZX0FE2b(<#a2A`sXpS2njq%@YcYH7dmXaxG#$PxO{y|p`nD0*{k>1ums_tpq|B4L)# z{z@cT;zsxBj~QXK*1!7FltMMFf6Sivf!o{w>B;sK(UzVR$rqvWZniD;biWvS#d2@s zU6!Y0th-Xzr^=XX(ZR2nsM8onMTA{JM{q<$;H7SnO99b0UXyb*uj z!^rWFguBo~L)ygNOr9&M(x#~?GjF@q|Hf1afb@YX^T#S#Jrzw zqvDH2g3i&}r-zG^yPhNwfs->gwm5tc$WT!Ohzc7sT3LI<;6Q$dxaxV7bDl! zrMWU#wW0WAZ;YRSqy4b6Czh-mp1)+FHZ$WE94sRJI#wdf5JEL7hPn+T?4_r)MDEA2&QcP?hR)?wk_PLLJxb7rjT z57MZHe>q!DBT~OfggaR#{TDmw+>0d0->LrsMreMhQkeKoScshj6awm1LXSU{miju> zZr{JnHfcnp0ED(CGm75fbKn=D-V=$F5{Y8_glqAW8oxXMDf$7sBJQcu7m( z8yT~Ct5->nw^YBXnYyPf2-%Ok;oo2f8LWTK!wRK=;*|whE}C! zdlZPO*W(BHt0rD`6wII8aqRWm^v+8&%)Ja@a{18Ym+x=LHyBw2)CK@j{YlR4VcH=a zFQz<1aTW*v`E~wKc{JqvlpgIpLCJofRfw)b(yLpU>3LFrvdFI$+D{n}@`&}}H7377 zkNd&0{_G##F4`o5&2*sAAIsa)Y0=s=QFM>{a2mD?a&Sy8tc`jZG0u>Tm?=+@O`)zD zA5XfGMAdtT`?=b7Z_(E%c$G0c8Nt-(Th`Q!&rJ>p@ZaS!H+Z$IF>?6f^{a?1U<2z? z67!xP6}4Rw9yVhaI4bJ`yt*!kq_7(RpO-Z6lOAH2@}e*mZho-8t!3=x9gQx*){DB8 zS*8-%qgS?RS^;)N12>=hxx(||@h$7Fq@j;XA!C{c;JM9ORh(c+GnZ)}myf@NjciBq za5lUH}_ue1;F$JQmvn%sNnns4apDpEsB&Co~d@^Bo2 zd?+k}pNxr+lD!aGJ-{ECD)}Kd(a*QMQ~tI8^+Y}+$S<+xAm&DBge?x4ps9i-Kbke~ zn5&m+DhX!>w{?2k!1U&x8&#Am343Gel7QIOd}eCv1?w$x))Q?38xf|9ZSry-Bz*XceUJo(tJ^xE=ec1fYRXf*8^bbevTN1 zQw}{bo>8C(u3XtH$gM5(Q4dq7%6WNJN{-j+jo4Xf<9$f1Tn;(*Kbc7mzqU|)rH~~( z%|ySThU*fkHcta{KbbZ0(l(e}oRwOQ(sCykfJH(?)*!&jnn zE_T48UrE1NsV0ukGMj-{5gjHj=t8+j`N*F-pSQ1pD{)W{^Kt05d@*DzbuQRAoKxL} z-T&jD%|jvwT4)~x544LD;mQq zd#y^~Q@-=PgV`o&b_lnja~Aab1t=E@Z>-I$ewYa2!Hp&7x=mlBJqY*L6;(Fz%R}}t zy73vWMwtxxCa8Yay{B1ci|V$HIChUE8NKvA6QJ)JdjC-LO{O#L`M~B@_Od405+&^e z8C)xc8Ho%btUWF8x(*1Uin~f}I%yv;Xz6BqV2mpw=;E@GviJ>Lq1@wxHNpbXP4;Wn zXVZ`eT+#Lo#PQ0(%;tF$dL0bL3`&sH)2;SDwR;PA{#$QOlDL=wU40`RJqF~QJ))r? z1k5+*!OQQ?#&0|421fY|tFt^dQ!z%2&fk=pTzQ$^XyT8@K~T#DO8*+}##+K!tM$-G zj~wt0vC4U6aM$^PAMkcgPGlYn?75?)i67v$G!myuRNIEah%wP3w<@@53f`|N3${cIho+O08+4`3BVGhoqSKo z`&RDpH?z07OMG%O#!<;zLkO0SFD`^A)ku^;1|L6Fx0u$30eH6)EHO0JP4{K+EvCsx zR+43v&DjcfKffUF*I{_x%61+8=Oqj;S?ek!E?89h+ll=10Q~_D3;}<*m$z}qLApwe zUcaJh`EKrKf7P;`XXJw+@hp$(EI{#5OK`etY00F|DDji~c>2HcjQ%I9BZrK%KB5E1 zR2NDS0EtGhpZ^~P1QJ0DhVievH~-s7_&>o$|H6&_$KU_YVdnqjE)3qpc9GbJBS*?c zNaK%@B`dJok#D?x->Q(O1#%Ma z(Ls>@rOYH%Y~#`zX7XtS)w%+Qoq;E|SZnddncs~<1tz$mGa?#yjeR8Kjc8Je+`7Om z4vUJo6MSD&X5HbU&D+~bymg0n-}pyG+0@-1fEdzlKrq%nR8B&7AH84Ot3Htce$lcu z53plztJp;Slnh@qb98;HJ?F+`I6@mew>XKynRG1|vMp<1oHqvM5znym4aK6nud04gqS= zkn+gc(pF?L1?C|v)!Np&SfBTC-Z@iVPAvR#7KTqqz%TEeTnl?(S$2#*+mYvqG42hLlaArAhh%pEe(|>ze)Kc5hG3p$kMm4r0-5)z!c7WE~i85V7x3}H2 z^UhYE$C}C_Y;VERjG;G!Q;@A&iH(b@xC*qtvfZ~HKlXN27(&GleEFEU-!ZpYa;rN! zVdE&n@GoS=VJtg3vuF`S=53wO*puwxtnI01*3Twlv$0oL*ZQA^#^|>ut!vzz$Tkq8 zCu6@XXZu1R0Hpr&E|%2ES#dv;_3HUP+D9Bw@aiM-P4J;!c)yD!wByvT=pK1j`Ng9h zra7-4u*`PH7@3&Lot^)(VltftLPqu)iD07IJB`G-Zxlq>%mEK&??J?8D{Bq|SL|+X zci``UHjM_pT(}%JIONHKtcsI+U~nB&h_jXZiMDE9UxAtSbU3vvPa^%enzN{=ri+wt zI(LW5sdQ2AQNn3 z7iu?3Id&eU9ELo2P~$Kv(b60e5|*EzpD2-KVx*ZpTE@0}Na!bmBX6n-peH8H zeDqE(^tt}wCxtR$-6>1mI$*QsG~jEG*?2O8f|2CFX}2M+KogZmyNx$W*{)XVJvhB` zrM#nF7ohW+rMuoH9U~R-6aSi`F@!`JJhWX4yo{#s-x=+!PCFfU- zB=W^Yqhobn*U)kr(NZDmxgM4=%1Bu+wE3}l9(kdEk$du;S=W@p^DTI@1DLvsN6fRd z`7}#pY*Bn*#~hMp^%p`FamJISuGt|pfzeEnsZY*Lq$P%`Z%By&3Vo6UAv;g^NJd=7 zpE~k%gn@Eivuem_Bj;CEV8x^~Aa+>-qU5h8_!9)2Ed%bAJO1j%;&3`gggp+L+~#ek zaQbnHXZazN|lf;`Qn9tqgD!I)GrCpo=Bcg-u&MaYq=(y0yM)hXS@jQ#K zS+^f;RCynt)waRasT6>pE$X-9rLc?fD$tpTtLw!w#yu%AD(~livfn6=yZtPLg-VF^ za-#&ftVh5T-mbd$W4$r1hX>F7+Ag$YMZ9y)XUTVe!Q1<3zK&uu-HlAz$JNq|y%P6O z+3;l^%FZF7mWYxYAIL_LBy9&YDg!qq&IIZwFv5el=q$I(AZomgOOr$~pZQ-?rL0AKjb<|9Je-3R`ffyRB!`U6MZf&Sp*|) z#y?3l?b-vZy18biv}tS8@iCgpAXVBe8ngXX^?bP}p@Z7uQa)XsG5bBJZr_8HBJ11T zNkI^rDu4CHL3o@xPq}<{J+aGRwy}OfB4(Yk%4fEr<`Tl9{+oNHF7F#nfEXRfFH$kFBe7C>l`C;`LCJMu-#nO|KDDh@nH0w*##D_n#l*=F^h zAK$qej_i3~M~R^@DFvzI-+T3qIw21vrtgGy9G~O(oRb7^wiz*}8nKk$Th^!lDkWGz z4g!mTe=c(15B7KZzsJS(L-ZkG5pXF*(4pZJPO|iG`?(0i*+OqSk>rH>m_J#XS?p!> znGi>#QBA4>+}hF_g((|Yb3Gx}SCYh=aYMj=*4Umi+34WLMct!;xLX2YH{U+sf6Jmz z1rcrsej7G^F$kVWDR}=)BzE%YFNM=+Pv#ms2fzzDluRjEq0=u2@%F}+({lO6WdCV zh>n$jL>>@-6KKepJ!gq13;bfLmayHb_oMETpgG*ink}PiFWsl$*+Jp2Lw91fp65z5 zpM4MOhqwIh1r!mPISLaoye~L-m3bjzt;2Kw>kWDSTUkfO;@x!{G|qWn3a#&&)uxv* zS_&4iJp$4%yf%yxD7yPpQ1_CV$HGq#Ak)mJ%19erN;h4z7#ah)F>zLU%V#la;dfqsM|19?H|xy z#=N6UqWB}lNN%{&v`AQ}f+b!*MH4PuboAXSYqi_@g2R^^V-o%{E!5fPt(>^uLbh%~ z9%D~r0Yg)ZhcbucD8)vmWA{z_-*utKLcDxa0teOmJ|(^ryM6)j+?Q8`Bg0>yuNypl z=}L&_%R|#GrXNny?mnt7jvZ(+2yUbm3lPX*Z5~;UY2u@5^7q%)v>pbDHHf9*jn1cV zFBybj=vZ&TWZci2GB?-6lw1bLvypF7;TDZ7)Z5806X`k(}ap?P3OAr z4@l?mdb2tqO8Gv@;a#<1^hY{r9``&pEf_>|60W2|kQQ*(2u~cG!UXK#QxR6U-5p1s z`L9{4hd1YLS2j7=W@fZp>y0zJDv_zC=3BOsp6?NM8!<*CnRDun$GvLd`3*)3$Y) zN@GQe^qcd=;L7)dPd9&P;vL~@oVHK5h5Sk1fbE5yTqcF%=}<{v-zFCdLek3_&6ErWFQ^Odi;x*e=cv7o{J^k4tHn)vgnI-0^HHpQ>vGbU ztwJ|)DihapLfN>cT8#MmL~`ze>@C`)Uzsz&vvI8#-Xyh`jvF3C`=xaY;0{(7?Pnx@ z`1qdf>h5CyNKcb+;GTPBY~IfIc9b;>RUE zLtc8$uIWfP-UuL{1ZBw3>oQ&9brN)n(T0b@T@cGxd*SW-=?Po7w4Cr4t3)^BfFPf+ z=2oF8r1Z7WQ47YzK4^wMP+y{Hsv^aaeZr9pBs=0AAN<7u7-#qSS477(TXl!`1zy32 z?wXVCgxqHdXPM=D&ua5F%mi{}vBG_hp#RJe@_;>t1u_xQf{JeQqx)Xk<`?x1zdunC z@IXG#&qo|R=9?kS9Di~n@Y+tDqlmMW#sniwq%j+QuQUzOMBjX(vuz9gQ2Hc4f&s%RvJeUB?e%x7!UFkT$tNG zSiI2@K3X1L+Nyl5pTYA|=4tA_;24MmCl%3-!gz9`7&(Pj-M@ z?>EFTE}w(lwlEU!beQ>g^f5z^|49_{s91v$`t7I)wf6!e(_0CM%m#A#2lgs|^b9sJeweNwL$cXHa+Ok(lu9_L6h#9n>P>Qn!)hF*v)72ml z&a=8|#Nl{Lvh`U?6U)if%^drpfFftjuP9gLn~wbh{3>zpD@R0!O8$#Kqbj2EU6O8P;+uWw?P2UE7J#JLnXecVVv~$Agf2$UOmao8aOAIow7&> z$88w&;C}Z{GH>%&-ji9mXx!3XO!bAw#_5_r113s`Pmmw7XQRri#<%$>CAbmj@41kD z?mY}1!s7GFk1zJUTd4QE8z8~qAfv-jq`!P4=CVLw&Cu;jtm;vru$F5si+G<66bPWL zoYwSd@Z==P;3b!BS510?a&sPp_=v3fPn**EfY1(o3q1%Dfa$HK`=>FJ=ymc;5|21$r~(U1??CYS z_!MnZR9j-syKlallg-c_F`t7MmdIq8mEW2YA_rWo+1@;-NysOOPY*9~?COL#h>FxA&bh_?Z2Qd7-{- zI66#1V69}n*4=EudvchP=id=vInyr#CeWWon*W5B8i$mN+QM{t;onF~o%br3yq&R{st&d{fWt6$g~9BkHXN1B8~?+a*>mO z#DdSuyUG*JqYI5qM~WNAYpxycQn*Ji9H!hi$CX>88={`;ASo8IlCA`V}gOE4^iA1?QUX`%-s zXiiv=eHY{+q22|^!Nb%%vz^t8&CZf(#rLl+*=RxT*y>)C-JM{dx+M~0Lk)L^QiC`= z0K;ey>l=uwfIDF|N7R z1|@OgHznb2D|h3`aq_2^50b$O&gjGhIECRC_kQ4EeUUuvs_%y0V#y{SC%$kaE`{wE zJE!o7>bQwe8}jXFH*Qci2$tlRFG@f9J?)VjRpXs%OxPrd336=f48Mm&k|~R&g7AB% zG%%_k9E{gh`$r?JvFV)4$wD0;_mu8=vazS2kJ$9 z2gX6;ur}eG6Kf*#07Zlt)VN)6DKzOydSHHJ_(IhbBH<>c`-E=DQ-DU|qHL)^up z_CQWL<&&^VYZvaur8QRP=$G@#BTw@z){Fwbvv|zh*cZQOIQ?3~kk+1kfY}947 z#tgV0jEdN`?!K3eKhqaUGmEfZIP{S#S)Gv$%S!ehk(b~|Uq{McLRUyWr!zC6j)F5k zoJ4~RjR3zIwCRHrjUB9WK3{RWM&i7-$mYQByDl_>6)YTmm`p?uj2-C)v?pDJVxUxy zQ#lkS>MstcKW=|!@%ES1LIK1kZ1meMM#4`7KEiD&N14nOMX8 zg(T;EdzgBoKaV4C%)kfYVu!NtHm6PYN zxp2p+)9?$uAUmHSt$%ocwy-KLoS;E2q)-uK{J0~|;`DkF>r5FtRrSLQ-&&pRlTHni z?XkIkYS7=tKagG~4iA9Ue!^vZX+k?27kH|+x1rN&WK8|1RuQ&CGjGQ(B&~i`o7Lpx z)}FlbptJ*=jLb-}(0)t^>GTO9;<2+nZo@N@XU5%74sbh-#{xng8VOW_93P*`ww7j5 z>zWTW#KL|Q&Q>AXw&(rbDz6ge#kMZHw;M^uFMPhT*JkuM>!-hs`6h8rN{C7gx(&$t z-#)CtBJPlL6}qqBHm<$F)7QG!EFn~iU;N2qqovmVsx~|>eyVrh%!UBvCas1V5Ol~j zL8aW^`ESD+yUo3RTbq^T{?PAk>(I52m>JhJ67NFc&zE(`;F}$ z0dIBI>UAM|+u@J^f}&u6cFPaW-27*)?5aW)5GVJ`)Bq_1q!+r=z+ujO*HHpTNU!j6 z6hz@x3((!kp`6!pC!}n?x_I1qn_8??N!{1-wr_X~C|gex{E3Gf2y~>MK&Vvjqs7EF zIXmMvL*Ivmpd)r%x^9@o?3yvqi-|f3p~;z4+eTtQ=LG~_NQMXI*h_23e}3L7o#RNu@y3PbJM$q?E~M|A!hz3T1ard|^i}b? zi)T8nN>l5ia&7j%T<|wBxc&=)RwY!BpJ_=@hACXce!P3%Vx$7DG5L^FLr9_~=0N5o zB=anFQ0OOpajql>34z7v9B0I&5cuxtC+41ui|UUjKOnf3J6beN?oIohO@G8$?UFe# zvSX@mar+l-3O^Z696wPYh5G^LsJ=rw8=``F5B^NNN_ah3Yusp0m_*XGlNQq%FX5k@ zcalnZC!Kqv7#my!u_hN$v;Z0;L0rakDKL{*=OQlG+wYj!eLqg!-$bWe`rX=;wt{Aq zceS{dQjWNL&VM6^7HL3Jg%n1%n(kGdTptB;TMq9A=`)p+mG24>osK%s4}_6fpYx|i7MqG z!O8)dVwJfCHPx-!=Gig*Cl}?SXI5gbWTzY`;i}D&)MrEWhLc{+FBaAGZLbxF7z(Cn z62mYWk2M^dt?j~VEaBcAO4;X8r!uCliL>?&R1P(D1Y-XMYzE^wd2)R8kbAPIU1}tf+%O&7xyksV3Nvv zN`In!MY1jMwIN;1pX4GId}gWdnn~mg5@(Klifm1Z@lD6qP*- z{g|s?)WEOHz(B7}+o5H)_Ngj;uja%QFtfb}6@Ft+RB3Cq39b%v4OCCgCHydSUq>uRP-W@<=!0%3e1;RY z$*&^9Pb@+&_Bz7xl|AB9-MWjH@-B7y>vGX`X1P98)K6)VxppB?pOKc``&TkygB-hu zx}S_>{vwsMiR-Aw#=SNCf=(U_bL#qRuY7mi3H#z^okXudw1uJgH@|-t^hRuZC-_=j z;`>_CH6*)g{L>jb^wpQi7iP!!68MCwOV4pEZ!uLC3H4GlK);sm!$7&ECD;F4OD|>r zslXC?^!T*lbkK^y&Gxpom@j&QJl1Tqp`7}AHm~XL*<=l5M#sV$PykawGq2}?Bm2Zr zY)(to3$7Ipa{}m2q-)mU{FoOK?%NvgDB?1 zTQ?RC+l_CkIkb7p^l^>KrcFaX#%r0GT?axIsy;_EIKeYLBD*Mz`rsq6+G9nUX@U!N`mNIUO*tf(z+Qqt6)ZHvIl)u~4)|Dig z@BB7Oy|Iu(rbnsef^=QHn&_K8QX9^l#BoL0kLSBODpBh^8D-q^E-5@Y+~?f#xN<<# zy+j%NkykcRs$BP317@Ps;6Qt~Uj=ODWXP3%k%FX4`1?J)(+|Cw&l=X>i@#{}fVaR% z!qa?Au(*txTNkHfq_KuGWm`5SBob5WBFttY#X<5R=kqo;CWqYxgikDu-UFBIXq~6} zHv$=1@rzpm*MF40hTjK4+49A};@1Y5%x{oZiF?1lQg^{oMZWGA zrg60sc?HF$9+J485N5l(-vHbzwlEA@Dn7b3=;P7+a@}MOAs)-uEBw6idpv=kSiZrQ zver@^r6oorZ21L|ucNVAwr(@!lq z-+lST>MA-v?~!cNz^A8;{=76#`nE$Gw=7JF=Q{Tbb21~P);Jl|j^q?etFw@{lr}Ae z&@;Nkd|cNck}bmc(Xm_hy3nPak2^b;EImXElo~~pT=nNiM={PNGJne9O@4dtD#a;_ zC0q)#Ft=nH{7Eo;!h>cPwtR~)PcGLK!ks2O!mGxpY~sGw^ZQe<5ioFE;wn)5AT=*#PWCNe~vTI{md@QTIuF+@2b$ zYpCb#8|9T?*J*rK+EUH;*&-p|v6!^(ab^BnE=+g$o*g9gQOySr~@#|^4t|3RC^bO!1 z&)W=J6kTsyn6P4~XlXY6(AI!icw*y$f&pGdc)yl0`8WnAqv&gq&WD5PBNuTjDWIUX?x=OA6iE4JAoxV%*Pv+t~UBD_M~y8P6g zB<&J(Ziy?JH*wpamxwW)g0X_wR$I0hxGO@4GG36La?sA#9ri$IWvH6_wRXl$FZ`9J zGkM*TOO@Ho<4ZgXOj|jJ%dd`IG$kVNX*8C5FX?)HNJF3^*+)^4IjsL^ zkLRCXNjnd6SuU)$p;Z3>*KbR^+S+g#Px*#Q8`0Q1D_^2sB`DoV1gVwnKRn5;C|8Ef zqtf-YIPN$q=2>8>pv3^IF6w6!<@Vit&J<3(vFP{2Esf0oEN#EBft@uChJ^sA+p2%BY z>!&9IW(2cCg!M}pJwwVGSY~;p?`hc}6|{nbALBduCJq!!WG0o_&})Qyc>Pva>>6u? zIgU_Xv!hgbX#ux$p}Y0^ru2(*_gfW3AeaB;AYI*hkDS~x>1z}t#^N_Hiyy*I92D05 zQhy?u9ai;S5D^z zM&l!c^Ek4s)zl^~PNB=F(4E;*^kzg&upLnIZCUusbLv% z8ItvgYU1K|ZJ{b|j21h&2JWsPOpKfT5XnT$F59&o|0A|Ku|VOIdiFY3lwGiyiLrN& z@w?+OLm3+%xje$BKH?IQ8!%rJgatv)SG%TaxtDruV?ou%X*;{_1cyMSvNVH76muAb z>HLYu$a3L&DpB*>eV!*j*{KkVuz^FW?VS-JV9}z0B#N%Swo{C7Oi;ML?NU1P(~KUqUwm>uatb6%7lE}*DOaef<4qb{ zVrr#n&CX(QETo*kr?yPYUaxvqL2sK*_r6190x_@<}6pidFp##=Zl6ao@ z_4976Z`-?)*rQn#L^4^<3&_;yJSlP(iD8vFY+RNh$Pg7?lbDF+m`T0}n|SO>XTOb| z;cUB{`-n%Qg-68?j*T^XH@Lf%MQW@=Z>_(<1Yl=d?>U^niEu(JfQC}L6;E_MfJadf zl1<`w@H>+{LH2Q(r;2%j@(sL0Nt-RN-K#%AE<^rU695-JCWFEba6lr=B9y)m_?zTF{>n3qDfT4=Um*RFS*{1G6MZDYA^LfA=H~7NHlHyemtB!R4cU7zf+J~ z5yWp}t0SnmA8xKRc8cYi%R8Vr$Ok$A~pl`x;;L zhx~b0GXFCxCy1q5G)g4hWDWuwORwIM`NK`j1``24x#>_7xqklVTEa!(>ge~z=KfxJ zC(o3Anx%7srtk7?2$kEtM_ERVyS029&HIR_j%wg{J8$c~dSOAO@f7K{_o<8K?{e3f z!;3{sm6>IDJ|9VbGCs>VqRMxW_{D8bR&SSaP;HGZZ1eJx@8@N?{q5UevkH8Ah6z$Hv;s9g3eLEuXCWT=zNC|EQzc zOwFk<7wf|lRjqb|#6OD0p{MVqT$Z}WCM9aEMpG0pj=YX&iXTMqMtGBRB2MAi*69X9 z-u>FM_IEHz5dV>y^JSq%{G8ekzI${dy|P9SY6$7W914?g`cioy8a_Cm)K}?DZCc@% zb8qrvi~7ooR=QpZ>6kDkDEDzIF1rtT3VF|OX|10E#hWH$+NZiN@m~(|v9+n&$o)zP zDY)ShYw79YC$9&~j}IKg;#+o^CYBKBTNZowaivA55KlFihjKmdMA*c=TgJZ9L2c)q&ir zV0n2O@7I6%Rt&ste=@>Z18@>p!6M2U={cE#g0!v0qg6kg3g}j+h-I4>9PIE`giU^q zeUql51T}kVa;FKj-?R`N;#GnW`5pK+LX^1@D9#{V(_1y6&6NF%u{-eOV-cn1z0+Tw zq%0%$jLuMPhe0G0@e+kIheyGoC=Ub;QD#s}gIMK?`+7<&EUx&~%5&G|?qoB=mHDfq zjKMNY%J;LJefMfzIF8u%tWUTiCrda$OMQ96^ihLF$Zy@BwndpOWoM# zh52LpuRqZ^9@QLbJ_%fZJu8qv7hsn~*l3B_Xu~NzU%AFZQZ4*nEMkPh_9;wXlzIW7i5k@ah>sXno`(e*oSwwG(K-H1QX1Z zpI&@kzhN`g457aMY81lqlcoyd`^P&)*Z`$Wa&q#glZ*HP!W~cq4_std7`F5@$plBp zd;NVzS@qJa*9F<8k!qH-u9mv$7*+;Ne#kL@(pI> zl@TW6y!5gIA#*5@;`xV&9fP9ZMsSco_mK1|blKMCmo9eh@DUYQ zhtHd&OZirY=v6^+=?hW#B+~3_o6q1puM&sPuP-zZd#jCW^*>XPq-AdXSz9FXL8jY*h&b>a^@Ui5}4fXOOsFVA?K5^eUEj^sh zA01VNxZt!g>4gzs)k@c@j!103d9i=) zsk{6tV_TD!`2m=vm?|wnSo0jBT{`pR{{+p)jVMM^_T&?kTQ5?%*rqtW=_ROA>3~zt)S^q0Q3a6@gU|m7f!; z)1LhkgGXMX%mV$R1d$G3F#uzD2#ZJHGT2y^&47!Gm;R^jL*C4%0rlN$A}nQ!PDYb7 z;DJ0wn5CicDXf3v10)u2i(Wp#Kc~_cxkW(_NqilfYXDBR#$ONc4)b|fzirO)aMSaO$s z1DcC(i-cbSa5uyf&E>z4Osz=NuVGX8w?EcT0K&?gzHERE9+IIHYZ-1tyK5n@lfnC9 zQYiqgVE~=n+Nn^f533qP^Xdp={((HD&r4p+EWuU*pX&lzYdqYfKNnkao3XgdUZDB| zBgm24DAdF;Ld+D21M9c+7otZDdmS35k*x99&gg>XHs#K0WF>0K3WXT z@1N`jP~{zx25}nPLvg_WgB$QuH~LY4MAuOtA@|DOGwxvH_KD26O2u=xyPV4(sl$?% zm6oId|41ABhUdwRb#QRI#=k2~&hBr-7LqxMOAdB?gP*QF;}jp3|2#@(DpJ5^4!lRLj-O-eAi zZ^&QWUFh99N+ZIguIKE0){TP`d(5}Sw_o$h4A}o2@VAFLLEg)l(CcP-b=%jThc1eim4Ph3h;oH5RS&AqFD6FZ#P#>1)2biK zOkNorD)c=vmVMgV{j}!GV4_4(mdOKc*ctE?6^>XVx{NYsKUo(t;F~m+U*i>0T&#+$ zTU5lt8HtGAW4h{CU=MHS@$O2@vzVC)i8zMn%Npho8i}k>prTcooubld0v%UAGE8`Q~ip%&h^lE@Mwo z&M@GL@VhZS|7x}{_gZ1s90PJI%V?4q0;_#nA zBv-_*yOZycmZhzEBratC8Z45FWQm_l(qxlQ)8h9WHt%kU+fuMNnlN;AdOc&nuzEwj zuCB}f_GXMW=>@!Vc<+-<6FD>R0zano4AL#Rl+Gp;*%wGGK=OJ_?bZ;I z%kKd!y11!kS@5J%1L2R6r)PG-(EQ>wdFm`eXv9no>0ezZ-`Cn-^L;z<;7zu8(B$RD z?3J4i# z#didl@mw^0LzAuNX$eCIT0Ou1RAF)Z4d1&);olWqLQeYe!bRZ^*HMjd^cdJ?qg>gR zVh;D=Z1+_rnZlYpgDHYGqEAck$aK9-`bSEmFk?<)M z@e`R)_p2TXPOk4~OI8l{ZPW&u&bOxCNQiW$n{r`$F02jBL-yI3XA?#jLa5))w0&Tl zU^se|ejh*l=of-0_6m-haa0gBStDLMPzo4|Tyu@yFF}eU0L`jT7k(cegE-Zrcqc*~ z-`G1YwH{YrX`?5;=6#CAQP&||@ov;}DL1nVc9Ponw4qty0EPPo3Ud4gTWZ7|+(|D2 zsyrF{DCJSaqslB#Il)YaKHC39)py5J{m1_vD|?ge7)NAg%Q`44g_J#xLP!W%bvnn$ z$|jV`2vL#jz4uCXj&&R)Lg%P->WuGwcYlw^?>_Ed{m}#Gb3X6)Yd*&tX6I)&V&kQ) zU#lk4pWd+mEx0EmZO>^mwpo#}PX$6ZjJ&vZKg3_;=Ill` z#`Cuf20TRDJzm_n)O|cu{G3benU6oul?s0AwW`57y4V~Dyg+5SKna-1$nOT}JHGPH zILGkw{2pSJmQ7CRBcyg$`y`0cj`mI6$mnL7Y1~+@9iXyzFKP;Ed<#TaoikV~|6O@N zXs=zqp%Ls^`8`Ghv~m5~DM1H!BwL=E zp3#=`wl&g$#F&{gX2{X4EL@7wqj^lcMrCoU0SGc| z!~ygiZ>I`A7%Q7bm5!ReC@wS8*l{j*FO91L+mWr5Bc$JZ-E3MuNye2mhd+~^j+ks> zqyzK%ppUYrd?NGEBZcr>A#ob)a$HdolpI}fjW>BPLhE^tJXGuyVJ-z^_p94LK0V>+>4VG?%4Y6UP^4U|iJ&yhjROEH2zJ zdOMhv76qYukvSd=Fl&M6Mx^=c&jB0EL-_?^A)=AIuKQFZVQ|;qZ|i|m-f~!0#S>94 z=^6-%ICI71lES&ZUfDEs=8x(_gk8(bz2{T6@Tx5z;w-mWX!NA)b|Jreb@qt1r(U8| zz(ojvJu^HN1r0)$;m$FMLQo^;g+>^qOAL!$2y0eU)&~zcD#FBb@%d*J?HsupF@2UE z5+nZMt`7D$Auw9ADEc(V*$pvx5IW)h^noMxdj{$p(Tg*+MnCet;ufA-b@PE!tz!9; z%hL+x5_0K5!)LgI?!3o7lYpT0!S3Ed7I?6@<_+uHzv5B-^srr&YSINr5f;pprfylS zbA}7P8OuWnJ<{M*8Y6tEAB$Uj4}L`w2Kw$>7n3x;it}_XG|6d@0v#{sGY1{Yb_F+# zz&tUg!`?K>;gXipvhjPMc>8_$5)Z1JM)4Lbf|m@L>6QZ0y@{B`E&GK+!JSCI{P1+f zr1POPz48C66KByRC@d5bCl#+o#lD8Qm+=4jU=7!< zpmf7;`NMpJm!bX2%-Cm2ebTG!`jDu@Y$H}uh!8LE`p8NPsi)hVy12_Ibk=KixAtmL zyiTv?55OZkY^E-6une-!T|X~fTySu_i>EuUNxDjMMJ=Nl^i@cPNJ)ZW7d~Td#05QQ zb+u}*-R!M|)k9rYi<@f&$CWJKcHf#wU*@OJDT9#zkc6n19vXdjk_q5z7>ZJgfIT;+ zGFhLb=Z>&Ar%xE2Dfry$YT|I7mTSl@dxZX047`MUwwn(Gh`e_#2UY~loAfQ&mH1Op z!s*RXKSvO^N%<x z$Oq$(VY}xHY&NGgDJrOuV*E8C-2Zzjpp!6#Ptz2$t?p|zO#6fdqfZ9aTuSq{@>~R2 z;)UzjD!M&s+(a{DQA((do(7Lng5HpN@p9New{*n3GJlhpnT1qGhun~`{}X46)0Rw% z=%_*8eAz+>DHl+vzi=)`4wC9=2JQ76r-pFjXEG4e{Jdts0sW(a! zv>8S0sl5)uVwB-rt?WC}B$xnG-D+Kn-{QH^Z+EM$q?(U=`*9Zn$b3;rl}b1OV71VBp|?>>l@aGJT9W2We06fKBho)(19{?eA< z{%*uZAEVwkq=bfUG^rc{i&iU=8PcM8jtV0tU2dF18o&AM>goi{eHrG662W;k#DzpVYxZYO7n5!us zpDKu0-u}FkD(PwaekO+V>&SW!3t(peuSUj^jSTo{qzZvqets87e} zK5Yr&c}fx|`3xbdfXbs@o?1oTx^uOZ`}=>8YrCFF%U`_wdISH1$ROFV0CDmC@LAZF z0gDJYg}MF?uTNGdBE!RPOT?*4D8Y){-jlA=T3o+n5UayasszjHE+SB$c#N*^DrzHe zy0#HcRlMFWC=7LX?|F7Ujz+|+R#K!~;R~10*qx9UG>9^xRUIr4ae_{B{=;3!QOZ&B z)LCsb4+=O4ldWC_2tUjz&(H0zrEjid@;IO4o1FOP!ddt6>lvX;;{X0!4WO+VBH4&< zaNos2r8XB8VAD2(ZU~&kIj#GD$^Rl$pUSGkl*-KJ^kzcNb|7M7b;m8d@)Iq`lSV-4 zL#xAE;0MyAt5M%hUJ)uY$PCyqebdsY$5Fvb?-h?EAZHV#RNJRucj^_)mIb8&J3y17 z?syE=GYkxr@@_cT<}?*#kTpw@CQJ9O$eIE*COrc^+t&Ixokt3?e_Mt4>U?2!3>kMV zArP9sfHFq^ic-+6ftt=DUtB~feMRFlK}%?H@3WH9PX2ji&d=SSJ|5Hw*d|TeGT+JJ zgg|T#Z1)h!XuNrFIviL7ju6z1h^e^e|3P?NzQ$V>jqW?LM`~P&2y*bOA9pb+zgT=W z@NG2YG>p1VR|p~B1;P6RYNAJo3a=cfiQE(ioftDGSsL_GTEFuNQ+#2v7;<8UQGGpa z{FvvUszlzFkTkCNv*QUjK!`DbuSI>Nhf7~XCF!0D$C$y|)YK}T%Q$uu-27;vmxac$ zG3S{7L78%?tlWfmxXJ-l{G1Un8u#+TfJ0&3{Cj1P^D7@%`aTG5sJ`*#QFR9-?>IIt z$54Zgvfier%q0Bi4Z)KT^=XBFeVybl!GK3EAXmuDdh*rnc`O1JQ# ziNT8U$kD+4(ZIO55xxi(Rg`#RzydlJym4Wb>g#9xtxWb zt9Ke58^5gl1qKPa1MI|zCJ}TU40e=kLj!UsSbD*sWEuPfvB0|d)9|yq`9B6kKTWYu z&iUlu%9Q`6!*adIK9(+xdhRm{;6ECFC6=A404hyGh7TK4r2Z66ZC5!nNvK&WNWYQJ z^R!Q*=N>}e>R}Imu=ZecBT?V@&8V_UlY#R1el#sTr{bq~bu2ozmO6B4_ewbXK|K0? za23E7WGNL&hA^P;`JCuwGV_x(@7qx%5MADu)}KB2q4QCefh22sp90WO@)8RsPd7$svB*iQPPfcCuQjO zubhxO>gImr0tnM#0LkWSx7LhifX!nRtSBI$)y+Ket=rm4&m@hrXQ6q=&=>zOyq~(nT&(m`} zr!x1E1W+?jR((f;Gy?c$P7_TJ!+TiW{Tu=!^@EPr)-8DxS!r^)C6d2;#j(oSu5SIF zF?kiEi@JCAqw4PvBj^a2(35!sPUf4sTN@FHk=hDuX+?J&9TI`*)#{^<9CK+Ix;J|& z5TAWF9n%b1QR7rPV(=EpyJ6|)%7ZsSCFVcql^#SJ%nncLD5uE3i@B6Ye^!DP)h0Xp z+=$2V2b>f^?T0o&L0Gu0B1HTT(ksq^=JxZD%30sXd|CXmMjK0i_vyBCd3yXivrq_# zNH)flbB@RFz-eQmRG{Eri<@dQ^_4ciC*MDbFZT;waams8Yfn2l2SfN~*hRQ@-gw%X zW~Jq{O?wvtfjUu#&~>th%-D@sc%wj~et5~q=9()`JXs@MinSwoEV?XlOZf{_)lN(O z?K|EaV=RQ3yaw_>2oGJPjh;zkBHh|Wb~#B^Mj)!<3H z&56%K>4LJyu~X6RA@}%c|I~@5N0!G&m~tM7Vi+kILPysGLD7Law@%=Yb4z6za0~D~ zarflN7v7O|XUb=RfU4ydkGaMtS1!UcWK-09@A#nT_In!oY#jPS@qsMKmdJZjjz?s6 zqOooI=SzqYmPbcHLAx<=`bi01=^##(-^%jzWTUl&IL?DQ->Epgm>(tRLRQD`4ixk6 zdk8i!iP*Px>iSG9Pw2Ub_&n11l$U!u1+mTEn8Qr}HXCnF=KX$>Po%u1E*8<_7kdb- z?n@dq5&j29hnfavxRW*9IsKSFksSNg!kUs#PC>XhnE7ZZ&fYX*L&)c(pnrNU!#&Gk zBWQMgI(kPJY7N6XmkSEX`HqbpaofY9Oirez_r&g+xu*FNb1Z}5=i*Tp{l zbQSqr667bvSlT$GPD2q4$;1n#Z@+d0~N2jIfPTJk5CC=y;J%Z9=y+Y7x=6Shv-~b=wia zhcpkA8T!?yin-mQ)l7($q>=g;+3X>ZweSRrZk-1+v;fBSSxG3Tw>!=3i{$&U7p zrqg(3{N5ZU4ga@M`NMPl+U;F$Z?}XZo(2cX z#7aeBupM#=3a4yMTIl19dKrJ-6*7slWq z)97q>-b3hyUl{bMy`_Vvg`%Nb_KTZ6FP?8;22Nhi2Onvkrz%C&1S#)CmvaYbXAQND zc{S_oI4y)m+1r)?BLM!t350h*5RV{0=DmQOPE+jkty=y@m0Eb;NHM?G@}lpdL36PZ zrng4_fj51%oLKxRByy)kRRU^=`dEy&DaPzxApXGr9xOVGhqa4m109}-ipa?Du)rO2 z_9uIH$)?h6x30^_D6?`g)4|XVT?#}B=(`;L<>HYKkPw{@7y4e3BOCh}b+q%+K3+a+oQ}PLwRb@@UH(8A zozMs#G>D2>DzJ4y_cyf1e+LC;>9y2A{DCi zgTu@G^BqDPVNWOXKULqc&8qYV@E#7`*x3$YIa*^txG3R_0}Gro1RYg^jHDFScJkwg z#|>=Uo930m=^OnWWQ%m@k}Q_-A}CQGm}VKL}pI1A{n(06LA1`B;&E)7hi%L~1i zx)vTDa_55yYg%W5&n=d++$N`g7US@)V*GW^6c2r(Gf945cz0mx_Cq?muK9y%SbLL~ zSZ~Wy3-_B@H_C2aXCJ@dCBa=t4J3<@+^F-qDvy6Y^q2!3-+P~i_C(Xi-yA#IVn;o- z#@J>xg35ulfw3qzn5U9lMIBXfQ(9LZ3H$ydpQ6u+1%od`kr|Db%44{e+3&CGOn%FGnCFgqtEuYKF82xw(m(#^ zjN*!#*%0ltA@Jg$iAcnj(zQX%!Dm<7d z8{Ch6^w&C8>_12vQu>|Se&k=bYs0b@!8>i?7jH>v=cwN8_GFJu#JGZS;y$9O_|V|Z zzJN>K^iU>W%KM#YH`8kO0YTruL0HIW-K%Z<+tTZv5D0D08_-A;WFl`6R8EZX0;$Nz zHxt$0)Xk%VmDUD=3$HhH6s}84Zvpr6d4W3(-1$KDlyQ^s;1|fW$CTQ};*285BH0ux z)ab5av7|vFz3X`QE?8bpvC=h3b|UX{PL;jUwR{Oty6l7Emf`~ik`3|#rHXuG3sg_t ztObL*jL9yuUobdK6vREh&WJE76SeH5e-TTol^thoM9mwyBu9srK9Qy*qy5mQ$bz60 z-25)?r%DXFMvO1fN~ti83=!XLJ@ys%dV0B3YI0mV)RK)BON$1(pUu^t=JNB7MA!Q^ zRYgP(?O}|%V4Uf`NEP#s)QzYK@DCHbm;Xi2 zu14s|D@)<$5e|~|%$=1*n@s*Kv$t&uc2biX^ zja7Z^hysV#%v;f{iYZ?#(&%WJ*xs6(l*BQ%f+q%R1;6PGsx0MQBbp(qWu$N^`RA4U z<79W2{elL>@RhEPX^~2d7`fTe_{SM$p=((oh`FbGM$i}FFI^6fLGXYazr1A&WLVtK zYrOPzSvvaV-N;j2Z%m)R*!;DDZ=6vnXe!~B$ffQ0JjBfZPlQ|ee^btWss?oyN)Pu< zH2|B@E$HLJPMyMcjS<=HR{B5qghI1c%5$E0&>611y%cBc!#GVnOkddNq1g?7t;d=`Qo{jV+*3WVBl6$OU{kpuyDl+t`8PNFR&e+cT2L= z?y4F6Ea7a1`y*oE_$zOyy2M%15HLT0a|HaxJo)XCo{;`x=>+fb5jpY5xvc4l_qbDi zl9ZOrxsHg;mY!S1%jgqxVm1jzY{7o-w-IzFj`SBPb}GHuH=Y!j@#vk=XB>^%I`=f~ zYJGL~joz{G5B%gq>cVLO4Ej)&;KF`s-OAbKad-hhl?E$Yjk%~N3Jpj)T>BKqw8c;A z0w>14zO$Lgz2!QF&CyZrG=5s{BMD;;L4f)K|i!@tr7D+o$kg#s3$l$hY5j zP31M}XaWWE&!N28q-8>NR@KaJQ?h1%b{L4*G|e{tJg&L(%0lvT{)*@&elu&!UWj+j z75^C{HVNJrVTcTIIdHVkkvW(2P2U9QBGQ`H&MZ=K&a*S7o$nxdsnIvhuN=I4K|A;& z9QFfIf8DnW62*A1LDD0toOlskeE~S(huhlnfJV!;)wNdt_sX}1|1jC|oADcgnw{}l z&ck0uv@cB|$Poe<+yldR+NeZ03(|ow)SOy9!Bu>a$}qJ$#w&U@JMiX{kq!Wf$sVP* zuYCk>syz5bz~0`#zeRclRyQYc75x%F3q-$d#Qm+Ra9)Bs4D>0C=Z(IPDw9_~ zi;1wvuMi5Gd6FlcW$~q}X9&?G+Y6;bD$SxaqH>2LBB&fo$WpI|Tsd2ZXM$4sRHY{; zTlMD8Y-3j94t1!o*NJ`=Ip;r>XV*EhG0FF0<2X$?Z3BMApFH~X>j9i7Gjeb=w>Q@> zx$x_}&#ja!DYvNyFqsrl#kYM~FHX_(V}ZQj01!PxQ^7opuhXkCinXuE@bQ-0!uD~E zWNUqAgLB0QQ@vZC=!564-gr7;W%-2 z*6hA2tb_K{t&H{Hv(NY~lRrq@h>Pj(kpQ8MRblPgNeS_azgN(2pk>i-|EkStzJm$Z znX-LT;N{!LlOwOHU|6J? zE8x^|)mUbPLBIDriw%qLta`YN;3#D4j?v`?VFXgUW`CIhN0A$(ItAyMXu z_NdXXFi?W)P?Gouy>KqPz5hY70Nm>5x06fc8d(?JqhTcv97=VB$}Vh?b<_uZFyfzXDPFvj%6q zfxw2Cyv2@XO)b9_O{acP##JGwP}RP^p=i#7-{59863>$i$PTk8r9CQaQIleRfv4$( zz%9X2%bg1V<|kLZV*i}Qnpq+%B>q?M0op&>i=;r7X`GjQ@E}=Vr2>{cULnTRoSA4x zL(ehjILp|wX#^EU8d17?yb%dSiu3xcV138s^zFV?^81pw@xO;RkfDjL0nsrH>IrKO{)a~;Nk}RPV&&3+jl!KV_r&)DT+R2FX^h|Ow-DPg?@yV zD4YVsR>h5`go?w0Lj%GsRwAq=i}=nZOELNCW24e?P~Q!2^P|E3Pr6KQ1@tK?hffG9!~L$@ z4ll&X_?6EoE8#pp8alo_h#M{!6%vZMIwrvZuT)|8|7(`Fxu$R#f3k7#&dN!$5ktLA zDbnd5N}31~fed(Whvg#r(oX+yx5&Q0oV_h+Z_5KbK^WV%zRmWu7e6{#I*njOiupwH z8ixdy`4h%O-k6(P-D9h+%mRqaW5BaP5>ierehnZj4He_f;IZA3gwtj3{ZVR5_nxX8 z$-E4Nv^@x&eYKA%P)MTyUF57s#Q3d3=? za|1%Ho)5g~h{=Ijq`jYn-}_MQt1!wyy-)r`bj1U)=Lc2D3;4sD%8@{XL!s4IWQX3w z*+$ZXBsP{InaTVR#Y@`^N?q0*Gyv#s4x&$Ov$l3mJ^WeV?cWMMICfos7Reht^v|^G zpy8H2ZbfRn;Zi*X-S5d z0_MRDfbyf%d6wWheAQofXddq8>U4M%v}2L7S?=mre*8P!mShRZV&X5wdz~~;iqU=o z3pa^Za9_ip-LqJ9Z7mWy*m-tf~QJ)LD7AAQH z7^r!#_0@FqEVCt#CnP0Kv5a5O(*d-+se2oE%a>5Q1w(1FbAZ*^_t;?C+gg)e=`WFc zyYb_h@=Tl(W^!Gw{wptpA(V;^aPteOgAv*Vr`5@qlE(t1FAf{%UCmb;kJy1_EN9Pe zCE8uj0Ddj$*jgb_Ps%3Tn-+8 zrSk{*U+#@fgq-3ht^IW*tbnj2-Q3YcX)-(`^0fSsO?ZHm5a0WsT6be6veepu z1Fa4&1|AH2Myb4%bVDAn&7rChg$PQW*Mf?yzO^U5MGc%N1b9a#ck-IkWJtD~@zif- z?64nm8QlD*$8umGx^7NZ%$mxg5q79Dvk^m$A%H=)&6I*`H|1GRfO_TLZvlm{&8U*( zA>P*LM*`);6~?Q*T$eJ8fcC26HgUpx>%8po_d_G-=YQ77l4LnbD^Q#k3%g2Sm}
    Rng^CE-1uOG`iyk@_k>3H=wmst$#5`@Oc-rjCg1BVqk7@i$+{R4QwLeyYk z`dU|d%V0RK+h|tFL+z_6krsW%z7Krt-G1GYpc#|6hQ2-hS!L7wFujKOV8lZLWGX43 zu+l(+IqC)s!5p=$zj)JIRJXAm*e?mO8MD!D{b716l2U%`m0q4>!hfCCN>v)dQC@tT zx<}>2tbPWIDAT+Ho8YkE(!IJdzQHTBwiRtdysQ0vNIqUgNiR zezuL$>Tm4~;egZ>yx8c!D(<>`6+~{|QueFl5T7mMuI!;D$b%T-6(9j{5s}iq-tpE? zc(8e9#Z5#j1^3iQRfO-E;2V~Rm(?_okQce=6EE-}_mL8mye0&1ROOY$Ut;&aA?Bb~ zh3>{%iFywNBzIXPnbdap8no9~w6Rt6SE zSZ=SFWeMk(Bm|M}Up8)7(iDjTt_!lTl+yf^OzDsdO7eYzrgN+)qoWJ>-5FfQJfxDTd`eE`&R z`CW*`nlp6Ry!^W3EKTbj@)$7CkxE*Drl+vc^;1TSF>w2zm z7LzX|zA-FLXAhCP`dhGsbHA3#!i(Mb47y0h7=6Yqn+yckN`G^JVx`G^V}H>45o_zh zYx|ar^QnW#)~DmVOeS%?hMZ%7?uqUL^Qz%ol1UWWndnG}w))hUtbcwbMlQbNg4aJQ}NYV50TRsZ<^V^ zQ220d3<69P4a31%j&$)@M=kfTbq@u@sHWulQE5Apm#cs1bZmcGjAd^L=NNL+@^M4S*PWRx=Z|QF8u0}VDzC>Q!ELU|kSEvS}Npf~5 z+pwVr6tpJL*d!_ZB<^7YFRol_4vPxB;XK~*4 z^9GRD>fxT+J>OPS4%q* z+8mh_bh9Pet8uQ$S;oi2K}Ksb_Gt#+Y0r|Pa$sBV>j|j#;=qckVK*M_Uz@PB!9oty zoT|v1uXDQ)=C^sPr8Op3^s$J1uXkKElSEMxJt~92BsJjc*3E?)*Ud~p+Q0sGKcgf3{a(F)!$ei*-JxLb zgZj&SCVaHRnu-UZrFScg?@&`b9idyFVXb)WqR8##qan|J z+n_;X^CQ}i$^xK=bTt^5Hm^jYlJ0}2Wh{gAmieP8_C~#TEj(%H;}JdR`p8G!d_?HU zeY`^~-pWDucf7-=DIe5F_oFM5=^C>-I&xne9}>7;YfbuGe{+?5W%Yu-`UD%H(QbIi zUC2V-xQ|}km@!V_FJOX%*z? zenRe5l#8A9qf>BP_*u)j7g>1J63Owh(x0IJTU&sm?gE=C?KCW7Q2uQdkd36vMX-U? z!Jtn`D$Co<#o6uXjPJRWGJZ9E!+!*CYA&wYM}tlFP@tYgOaA`x)_`!kDVv;kj=5G2 zgT4J2>;(iewucxjp~uF)IH4mGjGzmSWIloZo_fY%l_U>3Lgk1`a`2|Rfy}d?Vpi5` z5ka`FbhfujxW^GW(mR7Lrk5pISRq8X9BK%`0C+gU=tJ&JO28(1h+;lORh+ZKJ=DJ3 z)0AtI^5uB^G~3CHVp`SY#Y=y5R*hLqkyE5dKbNMYgCr7cI=s?`KBAp{TU~u{R5~Z> zj`zg~i9lbL$+!0tV0YP(J0LR=8auiu4c&q@k9e$hZu4ATcgg>p5`5P84hJ7cIUdyS zlC_MK3IL#98b;H0dTrLG31`i8N``x1p?oit(wbnct{bm5o`6*)II?%exL4{JeiJ!m z8-N3*?1u@96(l~gZF61AgEyyVKe}g^hVqDfix$JvZ1P*Fe&yqdqW{E*L*?ORfw4gN z8w>^xgpaLd^L56x3PKru(o9zW z5dZdv!_VbyL@}W;H_Uia=oskp!X>gsVCk(?b6SnJbr&mMj?x&7!G1!*UKBm8JqpIY ze7TU+H%+s=F?jOZ)=$?h)Kuz8)S8kPgOJ6jL>!ZjL>j1TT24IvZ(SotdD<-_JER14 zUuZh?SX_5fGTltnDOIm6kUmUs4C4b7fa|o@%ze=EvI@Tc5t}vrpMM$uTl|0X#^YTo ziz&(L8@&EkX53FDx_(`F$`R5=W??`qNV~NuQBP8M^G15sGr`GoBzy^Xi@*(K&KWrV zTJ7guT+g8a0HZ3m;42Rbv{53YJG>aTNu}kSALBmRV%^so2O8g;{}U_J$A64C0PP%y)Ra*V>Y8`F{6G2tpej$^4u{ z-D1R*EgKfU9@b7tbin4C!#69dbdPb@jxGhusyS~3cI^0a%mRM?rZVr4 zuI@vd82b&ZoEOOIH#So@IBv7wB>3OibYPTix8Rmm-6@Lq~4{!cInOQmpn0x4g_1f*jJi2j+RyIGAb|z{Y6zpo72* zdSuHQKQwRQqJh_-aCWy=^DJ{hJ1l@D=?Sv>BX5L24Aa<0BM3$KRVf z#X~&b5ctgJ#)@&hZZY0J2)QZ+I~mgJA|b=%P?Q$Hk(`d-3It0mjQ~zwqU^xjo?~Rl zh0tcJNQRLD>A*AEIbW``sTV)9&>0aL{VV(p_!to?K<{Qto!$tTo6b-o8-oovcM~Tm zjo*(tagy1N@L%f@(Pn>n>Mt>%w-|3(irZjg+75(CTk%q>H@!NCi*@1Q_@x0(#&7Ge znZ|$zt`C(z;;2qaQW+alB#kR))E=L*aHBOIeyGqA!!4YM#kCcS62r;P*eP^>wsmu6 zjgz}m^<>B3tuJTN&by^0tiAB|eH{(zW(0T_aSh0e-M~cNi)e{e7n{$~6-n2Me*O6{ z`-_D#pB`RYYk@7gcMnkt7er%A+D8)qgWRl5>=F4nP3s?dKQpuP0dFOOPWjpGJFsk- z8$3@U>^#rNtkTl5^0qMkkS9@@Wl1+z+--sb%Jw3&hNNFzE_@;{V-HiP3a#c(cR`G457F^+JU@%}H#?vqZ7YmFKL}0z{7E2NP9n-PhGc{`lrt;GV^}_4-2cA?NJO zZJ8JoQ6J&>>o$T>#al^nWRV)O8a^+*GNsAN9X4I&JmDM7x-s6oCp89}J^-mxa1gGUAA2A+@Z`g@Ao>7NWPVnO(S6#`D1fv68qCq%D4UalINl?$<}&p)N@05@wW}} z?8rd~A>ib7eDD31{@^;`iSAxo(%RVe)Kz`t!S;td@8AL^Psp3?BYNLc-Ehid;2a0g zkH!4})7Rg_+bx6H!)T>!<`?aZdzgpAjtCkGEU=5%;&aVk`eQ2JNN{nkPhIOpu`GVi zsHpK1en^>P(`Y%JH6hsx@maRI3!y85&^Fg-xI=B`&~d)8b8fnU12evg?jQ*jxN2C7>=j*t~cVA>nU@Uea3XXTh$H&}U`!_D@bu8q-76 znBmoVM9>{DC{hys?Wh2FS6`LDx~UNnN%;JA%c`xpwawKr}+7}q=Nhv#lF0GPk#C98CU|S zpVE4;jZjfVvg12vqF0}XX#Ay?wtNNePS>&b^dWV0lQ{~>pIYejY&rY8fWOW!kSG2U z|KLSVVv=yRLxR;BQ!1A~w8iOQ+pXGyWZEwX+8<|d+aiaa9^d7+wnSv=`-7h-pw#3J z>tepTL_B-zKNYFqvd&7p!8>^1>^8iuCd!>>z+t3#3W!iRo}8n+dJh;-#W5jZe?9b< z;dODuQlzfWuWHxwhP)D{p=_B|6SYwqGezUkG5()jnFHTx_aZ|OdmiF*)$ecUmF0`* zcx7{_=K0tf_YE8Gja>SJIDt@JlKn~QsOg9vPM~@ie5NhD%5DGJh~Lmw3&~_Sl{t$b zEtke)#F`tMr44Py`_U2TPN*RE#1K=lzQZ&40wX+UlD7ot9g>fLDh_>R!yL;z}4=2!z(BQ&R9SO8(HEz66v=p{0rH1Oa1$9o+a`O__rlusP7pVM?@b@Iu0P~Mo2alOucD0Sh-JBXlK z(GIY7{Rzd}Y!GFCkEq0YoB=O}T$my2A?+hpSvv+=r+Bw?JxS;p+vwXFQgx0R2 zYg}nGgjo46iq;dUq-$NBNt*&OliIwNifF&N%EG&1DU|J3;XLOBSNj}8=&`gAYHW{p zPg*_o6QW@$9;G-{Zdz(py)vHTY!t?3ZVX1O#|PS{tKw}`=7&_VXg*psRDyERXh!oi5V3(*iW0U!S>qWE5ij4;y~V%bZ8z85J1tpG_X>fyhNUSsV!Yw@ zhZuk-q{#OFb)K$CgOY0f?xyB^|*8Dh2(*M1HwzFW$mRbQvg+pP`&iER$<>ZFWost z+2r(1zEC52tS4<$ZO?y@U<6TiV}vvPe&xVJ$IYD&x86mOzDOL-}REi?z3c@xPi-a8qZm)u1dC9Jqz;bNE_rw z)|({I=PMmRU^(#Uh59xPnq!s229m|?jrx(^j*Z8{r?NSk-fHf27YFc9)UTAch$QFO zCOGGrK=UkM*`q?zgz0|J=G$Zl7|PGqAkDGG95&@w!ucXqF-qIoO1wI3QaM&B+z3XA zdhViEJwPBRcw_n7FuZZ(;726j>hi9P)VLy6ZrhrN4oSCaZO#vM((+}WF_w65#-W}~ z4BUhgVU>pw>+-YLoacUaso0RNpLx(wHJ}qGbJv~RIvOb%mf*&7SMAgl2r2=!ioo0T zr1Ufd4@aYT$%gp%xv0X2c}-Ibi&e5&qrU^XPHV*ZH9zuTmPr+VZnRqmz9TTX?lS=b zyUxH%Q+A*y*-&IoqvNIeCEc2)j|IChAE`n)pw+*dV`r&;*1-Y7AQ7Ks4(7&v#&B$L zPa|}sIP(g@YSX&olR(duIq$fH^R^#v)RX1-5H;C-ShO(Kfm!0J;NKz z?{GlqN1E*Nc-lJghgqOJW^sK-xgu}3H7+8n5iKvzM2p?7^E9J9run<20?G_p_e;OE z!KaeRcThY!$!btn9fLg?gB{z1Aq3HVMreQ6%w-$w$4qodl;f496}Ne*IjN?3FO|wZ zzN^ntqPX68o<3Lg_BvhsXS%<%&o8BUiLdF@ew-$r6*%Z22%Nkk!e9A8SuvO^2;&7i zva8#uq-%1KY(kvHXM3M@kuTU7qw~EIPyj({s3h}|h7Bc%8`$rx0mZCjXyxaq_-oCx z>szMv`5#YIS}dCP38TlO=PLSq{RsLP2^SI1};zscve2+1J**py8WMxw7 zUU}~_79FB6#$tVmFGo_;YYK8kB+J_RfRl?-^l$yX(Ik*ys2e3Pdu8v-w*S7F)3VtYK-!SG%$?ruWSc!Oot34RBPHsWVgsVi_K@WW8_?!P6+f z{inb`Kv6}Md!;d`9&d5{`xJegpckVAWdJ>cWq{V1*f!|pJ=A}t6=ikKYXPBp&w0SH3_kwxM3;b z6YvA<=KIc|?+~9R|5Q>l-fw>a5nmgH1KPw(>S8h8HZY(UN^8hL+%?9Z$$h&{zE=Ox zz9zh6sX4+jSXAbdyx`@GE4?h+Rkj@0cQd&u>1b@ZhYIl->0*@cWJU{X(JpY!7^Uf1}?_s(#G~4Ik%ABuIi(pYyoAw`s$B>T9g_n&9R@4j(*i3w> z{0wM22j-&<^^jBaIWKN%#Ro0+c`gg~@#A{xlE}QSCr7d)662sfY6S66Ha znpqo(RIJ_6u;hW6OVlCr^&yzPBT`leJpsW{CH&9N~_E{IyFHKT0d>~4nKv|tH+?P&2o1>>$7 z)KvF5e9OQzClPvqaIyVm?1F58d&2u07u7mdM?@`VvU-Ono$3T7XkHXjfS(RH^N|O@ zMB$_WWbg#!Lww$yT5zlz2VSqbxLG&PN?L9|^L9W=D==F~OT6E}jrKA^NwnxbW3nls zXR8{`+<5^{8YmJ^b!}+PSLiy0u`+j7-11L;3T*6d*2h5vvEUgzPWj3fVD)+SB*JMMt5Wr%=qhBW0;zhQ8CsVni82fdVrjio7r&w2S z-^BS>7CJeE?n@Lhn65}w*$M7g1r0DD3cCSfqfZfi8_tFd!owdUeF9$Z=_<#+TfXh6 zEp?)o$m}IQUFVgYuXJWAb;ejCH`b$o6a;=m2ky6#+S6G3*ia6s-7?n{2pxjnYP~9h zbG4!y3KnJOt!^EoeD)lWNTjQhiZOga$pukmU6DAAZ@~d^%x>u@>tPY#c4E0`uhEh( z7=-1%Ots`~+=1Sc zP??H*K=vWtz{hU|Cv?MXyvp{AdGFo&bnq(UMug-kp@QkWvOf{|FCZnk2{0@-q6muL z3trKl^wB+4*eiNUxTclmCHy+zFbA?mBc zn*77BN2fH>Aq+vf8wR4Jgh9th5s{XX7$Mytpdg`uh?I19i%1GeGeD#_VoVwPz8`<@ zd;Q+)a{nJ}25@S=gpjLU`+=XtDf^?zKke0J6lhYK6XV#;F>^)#&#Y zyAAcPq^+&*Qe{IK$}%)Zh|(PL)`7$^_HSC>S|pwV>ribo6d1NqAinw~_#N4oU`G?y znspkP-fuK2<_%<87Zwa^LXbO@5H1qXNL)S$b@Ynhf#96bPsIfM+hj$YPY$*`GW0K` zM?4E1_UrCNw6Y3l5)p5OQNZQH7hvRaYiY2Cq7*zM*6tSv2rX-hx9S=WEfH%(+f;e# z2t%cwtwgs39B|r4AYd;LIOcpLSJQdZ+!{HKlrQ#fz@>VD>O%)(b(^9mUSgjWJ<7?* zNPF%eJb19TV5tb#@7O@Hfl2ZoHp|%A+bs*fJS?}fd1H6@6q!A?;&UHRxQazd5}xy) z-W}LsD@ZbW`|IuKOeV(2UO~d{KUbYSz#y9gFI7Gu zvKCWs;VvDdH4xCv(-~`-Gco@2Z58^BwVS2FXJK8D#;^kvLDMB2qR+5fAbbYKS%s$s zsTChGTxe?D{lq|Tg^4e`CFbZBpCOoIzxO3Zn=)NjHF*>Bp$8DQ>qToQV~Itwb+1GE za-0RWgd{e%^mk5Re8$l8+y~1BffSJ@@_BEI<~dQucJg%*qt`G?&p@OGJmWp?7|tGxd}9ue$-1#w%kZ58(u0 z2Po7Pc?nyR*tm+xsSC(itZFgK^`WpRogZGbyytsMQKVh&8GWD0i%9s4*WoQ7m0-X6 z1kxT6i8+DVVWfKOGke2`hDr8&kWAI|FQ$nVycN73`#7Vw{(}&~*5C3p1Rh+WNym|{ z(beY$D}IK0X5Y^_n(`F6O@GV?Lw!;%DNKwsh|>xBAUvVb>DR=<{X&hw)AAlaj-XTx zuEMF#*Z6zrJeU&ap4l^0A9HD8q35dC<^pO64#MQyo8G((r0(-LmP_yO$#6w1TP=R- zuV;f=OH8NaAqU#y*v%;o(J^W#!TR;U+AdMV`e&Z{yQ}ZKTyHHL%mg#Y z=Xj_Q-zQZR-xhql*k$zn@flf}H#n>ASN;e2p?*2oP2F`>QaUe)W*&a+T!)}Jlxco_ zB9&4u?Vwt6BYjYtCJ^(R;Lj!~&zb0vWU@my_l>o~q=pYq+wWrAn}grH&mT3Gmj0qg zTtZzjtPL_)G;zPtlWD5JbM(49JHDl6b!AUE=J4wRW9M_pPgpYM#No zg5i&go^=Sw4OO!KDKJOMsp50#3?ZWu)qdTz7e3m5tc&gqrm19dh08j?X-^k_$=%A(RljmS?92 z@^XL7pX%xbKBQuK0$S~ND=_=pFW*4tI%uX}X0Y7Vb8r=vIyGyswHPU^dlfbnAEtkC zu+aR=I!;>GvAMf7idi(6Pcg&9wLbM6EI|t57=)f)^iN?gzMO9T5Tk}Yrs~+H58nFX z)cXglHNq6sV6L)Wi#iyit6boOMG!RGANX(ph5RemG%a@cUS@TYU#4**`wikneS+1- zn+O_K>dfMZFr2da2?aQKA^>FVpHqGvX;9$+natZ&vfm6#bFIo5K+$Q zoHdtswT6X*AHRE*t1cXLs*MdUnXGinOYKlzv>Hb|$*_L2rQ(>}={x#pbdYb$$p?~7 z;~uyPvb)?dZrF1)hz>z|B^~-3aCweHp$oUF&ho?pOM{aU{G|^wDE8bLs5^We` z`Mi=lO{)gb4+oS#RQmXr?ymhbH^a=d1tDA86!c_@p@-6`7mdA zbWwd-Y@I8rW`2KSp)}=Ky!Hd*vT{FpqU44NvEf$@qDlt6b=r@=As>I7vmQRUufa1^ zd+5Gd#Np1<#4_Y2=i!M7Y(UFby3p({2b*DXeYJ@72_R6P_n#tr341(2aC5g1`X=Dh zhpjx@Ds143yDxql$QEkxM!H0D-@QGUO}!=s*@R|{DP15EK&3DM+qcM;!E1g<^)|1{ zA5X*i+Cs)h54_Ao0$N`k%tqIK}IBvpifcwW4=Y|}IrG!(V%4(|pXh4C();^d~ zvQh*EF7W=?4N7gx940owh!WNv*z}I#{|pY=c{8|YnE%qF0N4>7E*q*o9exuB>%0t)sTp)>NsmgVluFfIF&0LT zRwXM9+!|q(CHQ@IA5Mt#1>Zti_(TsRoOdl$1|2h8j;PvApzmP!|76JA@ug#4&^1T3 z^#N-PF+}5=FVOtrzyHNFZ6JP@AOjOr6+*=GKkbHd&((-|+IwUUKuuIgwfCx|z}Y*x zQ;vp9#H;E;SKG^e@~7W9apwM^G--wJQl+@G#*eQxdG*8MSrE@&1)1NjdlRbTUeSv= z<}OebYuBc0$m00a%E8+GC>i)Z@B{<)PD&n}B{45IjC4ttMQDO{$RJKrQ)Et;;kXCb zq>uQou==&7P#uIwP}0EZQ}!&jA95?pl8eHIB1F0V3_Z}IPQOXH(-9nF+K8QW{Z)N9 z*P^w1Zi3e*8)LUXp2F>$k%wDe;&-7@W!Xj`8CIK|v@{cPy!o7NWvfR_ z|B;NqDaG^=M3N*NvuJen6J8Hmdm3h%8m;dXYNu;7SXX6=hBP!8gbryE zLDQv;!j_N-9K$}w23xl|;e?cL@u+lbFYHpWT^T7~6*W;Cd^^solWJJa?~+dj!gr__ zW{Vm|4i-=Y7zdOyo*ZlIp=y>JDCX~{HrE~Y^(3Q_#hc0D5sAqs2+;>UDA8KwpCNF= zOb1O4$s^c0X*aA5t$uOw!Dz2OTj_{EZ~p=jhQS~ks?H+a`Hbqk_V+Pq_wB1Pm?JE5 zt;wjebiaqOy0&EQ?1|N)NVWR9^(cyEl|Rc@x9ZbSB8pt}hawcL;g#bULWb=mc$!x8 z-y=M13t$$r{>)v?GE#}_nnZeZ8P>e7yxEMTJ0rYxVMI7RY~&__9;bo6-d3}Ju@HE% zu)ts_o7H_GFv`Ze3GpVCgg|`Y^Rx$lz_q-*D(rb&%RYuC>%G31vfApg)tu;|{o5Gx zJU4+_ziyaF8dsKTv`R{#bW(I3(RUN(Tno(hAE2Q*?cZMieS7RZk1-$TYV^3VJkFqc z&(znK#fNIj9s(gT#zGN^I0IzuPZ@6LT=`sTR?V8lVt$;otb8SNX9Yvf*kUz5^S@>! z0>bmre-lQCC&8&+6@>EuuLZ2DtH6g04`^0h!PkaK)l#nNnKG^h@{F!dRP z<~z2$27t{P3trt5peg(jZ)?zUU`FtDsW$l-AExwXXfY+oc>T8X^eQ-MB{RQHODONeM^ za7r{IeR5>I)zB68F z+gVaD94xK=QlT2*j~78TgF*Y+UNav*yzZJ-T#Ff>#;onp~P> z%zps*iqa|iDUfRBMPL4T&hh-4w8z+pR~aNM8Z5FU*`U1+CQv;BJ}262L^ASv{(E}-lmCO7=)oS>roreA)LDb{Hy?dokh`085PiSVf-g^Zc$1;s zK=9imhT;gC8JV|)1b>OOpYY29Zz#xDcDxUuQthPB{zb#aL_t=w7H!dY(=;!iF8(V6 zdj=Jo!TJ?!j`5!;ev^_)`(1IHZvjbIHo~C<`7a@Uz_7qR=%dx>GvK5q1yVEqgGkk3 z(j-|sYh@L3i}C~36<0KP-XKu=wCq8n;S25Z5ptEov;H0`_YIZ;-W1--;Rwi4+tVYi z{3z^`F~!Ls^0#c5B>W$?tj2+-_!X{T_#cFl4IJzKS4B!dy7n^7f!hc+6F#<;#hT^1 z3Pz5VmMCP8j}W19kSgT~ETxFOli*Svr{05ogXTLtluf#++Z#jOFd@tGvQ_mB!Zgyi-}|(wYW!>cxj?$)PuVTiJ!?*$;pbg;X>h4|3Ltb4%#XiZ2MAg5uN8b zRkf$Pe|t!n@|=-&KI~cO>rU)+?9; zQ?u3Zk*suw0-5>&TlB|2LdNN2fb?B&J1IcBk$Vya^g zsU14>M)chPd(^yhFLg%3P1^Haf8gq#umq1}{K8*+8J0i(gV_vJs{PG@F*c4)0)b+p zI07h^>1WYrFTshifJ>dCLIB;szMh77t)xn6@aUMI06>;R6g`<-)AMXHUfFiIB?y!E zD93C*{mt5V6m8wW&Cxkd;oOYZk>6iujn(egcavfPyGXnaI_f7p@qtB*oi6SbKlbyI zeFTj z(Q4@Z38^>G4CVxaH%=zHdrYpywQF@V|F>c1iH6lz=hV8RYGw{>-0QV30t09NoVMQ~ zPgRY%Oa8JBBJMi^O(rK#57)!)!mglTBJ zHTa1MVb4LlXTs$TL_g1l{8h!$Lk5AZ9Ub2P!t`e8< z6s2p83UO;2nrMCO^mV6Fup6lDba|uo=d|iu{Kfc617f^yWTF|ESW;qvoSO~FLkHIW zU11+6RsZo_n}nos4KEjVr!e zwF2BS-R0Rc?Uow9p^ZK->B)v<*Gtz6*QXP3f|dgi`a*!Y1hJ3&*h=_<|Nj2|Wwrix zS&l)Rt@vTD&>sJ085pWj)x?uFIc9QchlS|wNl(MQjmbQZ&i^1q<7;mb zv=^i+jRb$Wz!UHSjYz;*UWz=;z~(+Idc-b~aQy69%D@}y>+fg2`7mUCn}*Ivl6^R! zktEyO4mG_bK@2DqMO-h(z!}r1uDW*)5gw!6_~|8{d`$PMSN67(*LcHY#oDGPkoW&L zQI+Vwss9D`2ww8&-wW;i%k%rGe!@2b&sNTG&M7jqEeJKI9bt!o4Ql;s2AVxBA!!Ux zTYO@pj0;zfe(4pb(@8b!7FZg;W4?g(B4QErwLR;(Je<|fa4KPP!wuCIt5#Z=09rqH z2OsAP>A_Oc3`crhi;ulRKgNiucp7zZ%sjxus&<5yJ4mN5u>En?do04TYv#~n;-xv+ zc-@=oi^fdCq0(EOsS$5!AhmPwf2NFTXezK7t5Xd+&W)yA%PBeSItBUZc#=Uc_OE3n z@N7%1!#_d1M9Ev}wA*GQed?0m<;Z0Z6>fL6w|8Va#33o);(jo#%FkJ2{TUp`2~zW>C)2j9cCl-{OsF=i8bG16FyP;6&wv=W5ypjUWD zFq4#E3d$W^t9MVkw%fw~X6Af}W2vbnd34$u z>Q|z_S6q0=%Xww@r>ww<#wFUPhi5KpA)3Oa3Y4v~<8C*~y%fn`SH z5ixLzBhlvtGJ5q*IhHO>Ky>BOo~&=crb*6hAsoY9nw8EHf*}WQh{kziCqLnIXVitO zQyWz!Hu7uf77eJW9`{vQnyw>m>1StsB@0-tStTHY5dFF6Mj0o|?*Ba7KcxA)`fE`; zQFp63+ZL+TlpSdkOBSWVe&u@gE}RXg{ZC`!5K7S0hFUo>DBUx^2>fL4C(y`px}w(b zC0gLRX$EDc)TxjnlxRCI4#!-J87tfi$+vYfs86)y)?GK0CezJGqmSBn{VA0I+EO8y zhmS!x!;%*NgNz*i?C4#)u4?Lat+}?wdikotqmQ1=!?Uc}^)BqW-Zsn9w;5gH$UF$% z4CZW1fL$}cD#u^f`S8SiwyPyUHh)ZB18Wds&{|$-Cg6Qtar$-Ej~B?E#?oz;-^yXa z0&!0Xb|+cEDO-k6gEt4_nDPN}YSp{I)2d7{(9Ne6IbF~ zUJ`=ak;A^aWM_)Be5C*+`N;v2J%Mc=g!6;(E$AwbVNBI z>%+-x7Oj%&PB-YvOaEtU4YWH;PzIWoGEm8C)>?36AnRnGA@PJK0(c$Oc8lUzP&$Jv z`tvFxlT1%AR>>%>B4H&@2XlYJ!Gd=QOPNMeS$3VuKv~XwDtH)!P8)^`yQ+@hHTxr$2bDWZTPKs(Ea}Sj-ffWIsFCHJ4R&1Uw+8)sBra-TKxSb$O zO8icUxz}DQ+g>|}HxeF~Txpch#xhVZ!>)B>JDfd23PAt?>jc;3+3A zt;UfPj)n{u`ZLd`1EsANQ`Ffy`|JkG5RF~US6`;Lm!7@lGv2Msr`|Lg)G2)`B zMb~);IL6^EFuii7qumSSYyA(xV`;M&PKG9Z<3Qiq)-ikZK2c6l{Fsr% zBy5ryTlD`*Moef3mLC1fD+W&GGTT_Ka_@6vYmQ~L*^zaE=agQiGpk7nC4{I=FrD-Q z$GDHV#4e#H2UlUJ;~}}a@0+OOIB~gsJ^KjDQSNWLn@@~$DCb&aT9ddnmvXSoK;n^S5_?eqNQkW;v(2$6W1h z>Fsxt^(>+l#Z5IADI5oB^i&tl6fW6AAERjx2`=TZ63CXs!*w9N3-aXau=c)VELQNf z`NQ1@q-!!?)*79~4+u4iX+Q4&`BxP)u>3<9ShMk<%2GAMiHOGo6rG;emlCIv{ndNZ5e1Sbg6N33bp%1|sYB*Ev z)YXORP6_!qKs*>(+{$4`t)1f(UuT&Q;9cF&A6WYhxm*yDTGUi(B-m(OUkH} zlk!h4rXM?or2Y9H1mle^A4;RN)(Wh8@_9+@y$zuw!^-l|Q?6I^tvbJssWLJl|Kid- z6Y(Yo;JJ$knNXSK@nIF z&#&-S-r{uItJide_76Xc-}dAf<-c&&y%$2tWYV9A(0j z%EPT{y;)0%duYvNDa%7+vyW-eq_k`eQF4s$?U!n-in0%+&lc(BU56mlSyrjgAZyUL zbnSQG{6%xU>Vsa)jA#NMLsfE_FPkoKso`h-JAhu;Xc^>?07fvqOV&NKYtY5Vb-`iJ z3$=2Zn|caSi{rw{*_LK^gpAZrQl&|P%Pw%t4VeC-2d$kA-y16pf=dHaSDugB@@qq@ z9EwD5CksKhE}#o%jF;>NYJ$)zGz0a=^m@viy%d61_=`}OtsLTiyX9?6O2M+vH3 z+d7m+WDpawliEumEdTI}wL?MFwY(f?= z^~ydb#3m5c!!6Db$8t_eh?U4FUKyv+gEdCE53eztby(Ea;3P`z_M|kWL^w08Z_n`6 z*?C()yfsI(;9pAt;sIac3zQabE9Gixz?mQ_ZKi*MVhq{Gt`klEj1<5=co0 z0ri8c*1GkTgG}22_ED7Cf)U;tWM{^2@53hq4<$eu(gYrKJm{T~LJE}Jul$nV-R$!c z&-0%cz6umK8+IbdD1F6Xbvq_8nPesM?c*LpP;c^KO@f0(i_sgd@p5U2JTCcWD9i`Z zp^2*71j+_L?W~mN1+knndYZ}aw9201txiQOJR8ibF9=>zGd8u#@uO3 zydKm&%RG0Y8ZyV&%3tXLIutGV^pw&E>7!Bp#D8tzdX_pRIvT%_XCBX@-NEjeQU3Ji z$#$`SBCYaDNwP{hM0AUU%UF?>jOV-A-P<2avQ5Z0(@sSGJi`JSgT(o&8nGd_z>hjS z$}M%%3sw!7k2a1)2n#KLs>sx^{YC~oSNm5b=F_kuJTakK0Lo_g6RK1``LR2<@`v5G zX7n2?l_1%Kt~2njoC6j_?89&&}_E@~&Pwuwt0f5cjSh-~ql8MT3w zh(u042;wq=)UI_%V1>MfeZUBG(FRmx0)<6_sl32ml;3e{Af9H6k*cU6S>iQ)Gs(BF z5ra;*;1oeX#`^uzS_fHhbJKE8cE#9;<$f^caX%|Dngx^{Cyp$BUdi9VeZxY8|DLfv zZYkr%W-ogAgy+vEto7=w4PGS_-kNBC_DV!ZwS3M5fA^>n zJDgv%D5QGL`vf3l^{uVMS+Ng_siV9JaHyy1#OUU15UH1!6NzvScIO|EePH#amh+>2 zy>#}d3y2cvW3AnH0+|mRMRY;N7)fkM8MbiPL;mlbo^}h@2WeZjtCOmN3_^r)Tbc)Pp;| zT~*ikO6W%)JR;Olto>z#z>2^Qvwsz|<8vaMBFi)z?Jp}8S_FHXG3K2kET!firF!T6=#d{rPn$&UJQkMfag&KN49AI4tLs(mG2*7 zf<@U)mvaM+?7h=9(%Ty|rA(ML?MZck6Ok_2-`7_~>U?`}7Cd>7ATprEjXC=wp|R5a z=XyjIgPt&9p8`Wb))a%0qUx`%WJ29ZMO+t1Zr}?WnzW$|V?H%L3`Qg1Sc*%fpf%)3 zjK@aX_cIlX?k~A)QWT!9sVlZoCw;0DOTIG6&cOfwuT>)6SPb^>L^iDc`f$8vH}^?0 zTS=Js%Kcs?VSN%HIO-z~kh{_gg`qjSrBFPvI9pjN_nU1Wey+*tFU`yN0k0i-QVp>P zZ<*g<>;1m;Tf(+cC7`9Lh7TAI8AF2h0UN>y(tO<8NzMz-Ka>a54+(TezWOr}ec`Ri z8KuadkpW-6Qpk8~Lbff4qu=RQzlkb3sQqM!v=!dG$G({W%IW=o!Q!x$(Mf1;t8 z6@4HC^=r%+z7}3Ht2C(}24T8?_kUi9|EB*Jrh$Bn-zkj430nbuG5FP|!I#?LD#(QD zAbC;?70_mH(D1%?+uC?Jg91|zx%{1mu82Jbx$(PD1J3BDx&4w}Z#1B{kUd7=5QQ~3 z0Sy$%xwWC+P@Zf2nr%<+gS^vwS>XlL-o-D83oyr&bvXgT?~KfO#EnmHT|V4XSWG(< zLtE|KFF3lXEeJ ziFTz3BxBj;N7e6QPLgbkIHFkMD3GDPYcz=*=^r0|E&8m}sl?dFcc1gV&PQvR(c_3u zpP;~Z6he^s!VKSo1QZh1kS`a6G3KOj4#qEHeyP}R2dmsmF5c``Hb=QJWoq&*FLmfZ zot4`&DYL=i^Dwz;^w}=_6ZjxzVdjszv2RfC7mx7P@vse|S#d(u};(#Xtr-vB={Ywro*91?ejyShW5nuVU8OHb@C5W&6Q3HyDd4R)MR#x zID?UG@ek*=?jK)7vF6kQ8cYHt++A($SB>fdWpgJ+Vzpw(WZfYQZ&O0lN9cy8AjF=g zKLQ5tai+xzo?k?UlZD!&?5fdZx(fo>GUs;9_NgShMaxop{p*HD{1P|ZqpZ}5NlXVq z*CclC{CiXoY;0cyhtuLnG}*p22-&OoVEmSrpI4{-V6pfP2sD@D#h;>0R7no*8}Tq~`qSyZed8zf!Ec0O`*jS%}`_Y5S@1tl4NKPu4@8FCGz zKB|jiYP0wWxG^?#mp;}3jSovA-mwjo3 zF)xG4LIjwWkB@HvcP=(6C95CC!G4@kGA*HxXB;1GvXK$-zg_nMjxGCv7T65b@j!dh zt}+PZeO@j7SF-Hz$jd&;beZ76{bk8f-sIG~K|O~usLeZ99{_Z}#u5sf!Wp}`xYb&h z415wzP|MSE@XqD``rJ{qYw9o~Qovl0IGJaw98*3g$8o%9jU1XU@fVxh8M2zYyXCu8 zYM%LC0@nYD@Z)bL2%z6D)7HhCs`^0V(tK6?lI2uu+g;iDRy%YG8v>CuGDJvE z8;W&S$aTERW@UC!vD8%}rl0!Dy_Sve?*T_#C)Rv@X}m2D>7K%2l&Qy+)X`ymL-eb! z9&{^fZi;N1d;5{|EQ5DGHK{hCCS9`Ygb=xsiqudj z1!@-9VIZylgK#~O+$&=3ye8JN#OLNXbAC%gqJZCiP?JiFF5{b|^cOO_A24u@QRauY zPew3zq^yjgw*rAzAKPuE`_Pb^VQGygK)*xF2016}$z3D0+q-TjF)epZ`Ioi(IB;a; zD-?9j6trZj6uI>!J(!pi@T1ePLx8G*uO#czAk-a^1gF70EVcMh;_Bt)4}z?bCxzdM zpE^9bk?&g{p=)1eP zOj%16%PFhRW+G!IUkTQwI2R59H@t#sFuux0xB@Y?9NMnOoiFGqTlm69{c(?Hk2~9f zmqqlo>78>Ouk(I}q*L_%_2T}H1&s$ud1qVndTqqP`|KHxB08l^NnJfkt5I{`9Wpmn z*AGF(wDS`@NZ7y{9ekwpx4O{GxyVn=TO`#7jQV}rdgRF#>x9Of1h!2GdU$=|yN@6| z7)~iJdzm2iL0x3p=^7Q2NB!y$M~ykBbe$R39MG~Q;#bJ@Az;Zmg@H~AHZb*%7DfQyd@a_f~KPCPB{4g-YDBHQ}pQmJp` z84wP+Vg|bY8P=sGF8p;F%xN*hvwvq-;tSBxzCIaMrpvp4Hkui(GWYZu_6@=p4XLiNl5N~h4h9}jaiq?@w~spxsYb%d z(P~Q;L9T=B0o8t-kIfQij@VY5h;KR8)!oqKWlSd2I1>4jd)0Mb9eDpL2j<-BCGo2@ z@p!;o=4%?i^q8pkk1AhDaQM`GVLc(>SA7fEo_}{n_uMm>2Q2pSmtfx0qa)r|CCqD# zi}xtyl`A%R!NT?B_1Xf&)3xX#R1d;+k_{)s-e|?RFF&np_f3Xc+H#U>ku4Z0Jxvn8 zCS3ZNI6~z2CoC|1G!F|Y#}5`UdWwoen={sDH^dlUIcDw2*R^kP9{M|0DN91RL9_}s zzZB6YjI;f?H=h?;#MaFt8^|z1e|@nkYnkqd;AHcJPG!BIvei@3Cc+n1)}?;j52xs4 zD-H@Ad~yDv$m4_M(vQ17VX6VEfj{3{oTv^HK_R+q!|qt~C^zQ$CvQvpBF!FX5_3jL zrD_uKTs^%q0)0jKK6`2f(l}}y5vbPd1#t_6=!q!G7(%?1DZVGHMTa-2;fgp@t z&?e(JkZm{P`v0ECe#Z2&nSI$4iDq@w5aoZolu6EZrk;&k!0CfuX^Bl7N78h8Iib^f z=jSJbMMk(QD)d}lWQ-2IGK3_WR7n#MjBfJJbS~g2L_V%Ah5vlPT-`O8aW}YZ&*y9W zkl7;x-wjr3Pl9Xm_BZ0Y#MaekLinU3Bm7$@lxlBX;H{9B87CDchIcj6F z1Xt|kIL5v6p7o-WP_iT>P5R&DycQgE(|im+!az*1O))aD73`!QupEtJMRwDp$Z!lT zgY9W4veuw3ogEV|+#5yC52n)} zQ)H)t)&K4~m$JQ!yigdMDx|_gwN41^uEnkkOTkwM#g=t=iGCSQ#Ey;)yZpWSq8ST4X6?T(o@@AfRoU)%yzfPjpA1@>}f*DT7yW3r(bgN z=EvG!0IgWhwG(g2exD20OW$RhYR*GxeENvP-k#91lGV=iygOd}i#avE%IOwe**zWe z@TZNHGfUNdT~b2jCTaRAibjW3sTK*g6Ti=bw&@%U4?UMQ;k^olDp*>SS8pCX?W}-d zf5X3ISIoaU%f;eaokI{gnlFE9_UGChv{sS=Q#sO{xb z(M%c9%pvAP?E{Jd=}PGe)rb*R;)O{#%II=`AGRd3kf!yW3ec})a465UvdVAu;(u3vOFKe4py_1bbhTEw?ff&w0NoKVi?y~j8v7#7ICU0(`fhuYvERy zI1L4C!u&8K#6R%zA&;OaxA45f^(({q@s6iwY`x#4-cp~f6JhFJn5)%dYy0txQ6fc5``2H{%1BwP zVvs7&1BjrvGF~x|IoQ(e!Y5$3>n$c-S(`19jl%^KrQDpS4;V%qbuY4F{cYA|gg**a zP?9--5X}}%ba`~BsVFcQw*5R0ZGnt7o8Px#qV}c&lV6l3=SA&KImkxU=1EmfOOfl+ z>Q|RVoh^!eUp}>xJ6##@<;xxy%HV6xCW3ee)RD14T#0tlu+X9N!_qgl z5QirA>ruB_DsTOXCK1+QBX$CtMo|bvpDHfw@+R&MnrsjeuMV9B;DUj^$IZnhgVR!t z%t`7KAUI^YLuYUDi7ATW->hiko`Sv_oDU~OUt4Ktd@xYHQZLO=RmJB>&Y!Ni#XUz* zAW?O?JAS>ScK8EPJ-Muqxl8h zADLnGmeXR*`Wh*nGq*iFdT)KB(N<2xQ`d@f|Jpcmc^vHQ`1Qb7PY{}wyrbUqw-Vi8 zGhYK+&xR2?w)V8XaYl2O5yfVYJ}J(_-bu;&G7Ne2us;IQ3n2HikrLb;Y*_8uhwS2@ zB7?bF8-~wTt9`PiLSr}R$LYufQ-y8)Mpj8sLo*gYrSZZu!;jC)qUQKUA1}&%4Xl4InCBvz z(+0D<3iZ0j)n0igu(sr~{IIwGM4|GC!HA3v8aW2+I1xzgxQ(K;BC<~Sx*^q|Jdk~d zDPQP?TneTh@`k*ocKMvv^oA%_*uvQtB@<%hw9ohmq6y z{3-gA@|vWAnY4f5Txlj1}!o>4Hmalk1_*9m%wRL!?pUgLbS zW?S(VV3u@UM-1G>s-Tkkd8lFlBUX;%`|Hvb+5U4oAz6s2^=7m6)lW0_F>>hV}XgmLhjFE&}7?GeoASJtX@bs^EorbS9xR*1_|;#Y&@n5%Dp#p zlukIcVy8tM!L2ELKu+u{G7R=<X~h z(aqE$NU9se*=)Sdjy+ibi@D#`eN~B<829tOV4!7wiuMc;@8j*r1VBJ0p>T|eBq4yO zAYpNr^6q~Ri*Sq~nB3+>e9%Ty$INwEOIBgAmBITqLXarpPXXQPC&}DGgj^zc3Mb^D z)}!Xxd;nzZurQQ96Fawj{FaoLM_l5DRfej@znTAqCy4-3x44g)$NF9JE@DIQj9kM1 zL9FLg^=GdN7YDT|2cEHf%~&-vg~$|BN*n17lLX}+pus$I7prxt8EQ7u&9Dn*fkrMJ zYs;9o7k2mRRHtubjeFfn65ler5zk9UFzl-y3To0oCwoG#YqhPKnMUv?Uc%So4kplBh(5V)*`S-J zeO4)H@tr!AKjGj_bEH&D@v?uaxgDElJYF&Ih9!PsL&+0TW)f^#@r?Wj!3vZrucS;Y z99&fOC#c?=F{oI!8DXWNcpOt}p<7J-`Ul;>Hi&?}LIY$$D*PL~u_-Q2t|!5cwq>ED zzVTtXQN4|&Zccs7^ECgebfsi!X6-K(f{^Ra01srwQ=fJS`?%WY;DJgTax_S|E+_rF zB}1Dko%9bh;XaF+y9(QDYkU6D!k`gSPt5O;M(!_S{0*tj^AyblN&( zO^~3tzT+RK7;h{IQY4KfR1o!u{bqw7?a$TU%Qh>cd{Dsgxa5wnL~uNw>4kv6%h~6E z7XjP|&s#(jAjR`H6b|~(Jb31ShT^Ar_&SnCa|J`U?Yq>3KJCyI5yG$LS&jCDq9&mG>7On_iVQ>f*p%QgN5{vC<8GgW~v-`cE-jklvxNg!MHNEPUf4rg4%OD|0VS zcj_$2jLJZl<(ZFmSDde&z?%fouLcYjS+*`rK<)nrG<*kYASC@6`sJz`He=l&%)qWK zg+FgkCBH-awYKC3O%GN512T@p9AUz&h~hA>jC;kjSqJJA_^(CvX??X^HoNL{Ei*Xw zc@~BCGZmMfkntxD_UN8PZu5-L&QHVEPei>*k?yE_p`crivNjORL%5~{Gtv3V1^4b> zO^2n;^t_i?jO*vVfxbA{Mf3OBF*GT(e`_gTTeQ+(y^-VQU>{FKSDY0gtux<~Zj60Q zyJ!Yq8A0U^z1*_DMv05W#`|E=e@w`ZM-B_nxfq_AZVDU}q;aoP;2gjtLAOY?MrE=4 z`|tCd917w)Y0^seQaa7L8SkmJ=qMh3W=Bo_NZN#ftW&eAcd}wU8#?`5T7BW7b9`-a zhCMwU`S(*EpN!Y@jrl1iYTvaNASY+Di5TwO$Lw6?t1CZ6G%Ota1avaxT+@VBUpJ43 zjA`?%y~6CGt)ABtQ*ITH1DzwA>yVGUguWW;w4mF^fw!rFu-cZ$PD7AP_yUsv!fKaS zY*+TI{D)rB4`G7&AZmk8s)kBU=zRyBBbcNhJhwrM&V^XdN8tj|UAf8XEAW^`!k{Kk5Y!U4GKlV!92Y=xrc4578jreH* zxT5kEF2Ut)Es`7|`>#a>n=&6XpLpadCSY!veAA}mOpXW|&HypV6z*QtKbFTY)%N9g znb0!2e+z~NpAZpRz6+UW5aN8r8_Q`Q4F?o#V=mC6txq1b0d6-Z#is*3RnM${lF(wt<$957dZYhx%oR`Eu|C z6TzS9OB3*^bns07miXxwHd334_rhjzZU z51DNRM~79n)hGE!qMPQrpd7@Cem!C860WKdxxi4l^Tt2WJJeUQTb!%syF|TTRZP*- zs+R5x-hcEuDKFukOKhY(y`2qsVve_|^Vy^NI}28|8dHr|A?&|NYuZSQEaE2g9AvsL{h=>FrQ?S%j%X8Nb-$=v3LV>*$# zJJ^A?^>KDU@U@LcN-kBd#nkVs{CH~36gO_~%cY_6E{R_6>(FA@C7gEkb|E4Sg)@KH zB109Na+D0y7V7ANct|mj9y^AIgU)O=As3g0*&JWk!5T-m%X<^X5O=<)3tu#$yYRSf zj84w8CH_Wpx{8V5Za)eJ$BiKBCteH#?ye~oyD%ukOI3Xv!nrB7zR5#F`a?nhiTk+v z@hCJ)gJJ^e)E3@9w%yW4psPs*v&DKf{>$Y zVj+qPMFg)YH3Wl~4%lMR_1X_toR_XYU$gWM4&-=esJwJ|AF-^ZQ1Bt0Mmz(J2A9Fq z8Sm4)N6sSY@Cq2*vD24dDXfnV6&2MlX<&y}&pqcdrq(`8u}@No{9rD!MN1=2_xmDt zhs-;&-qe$pfRrNoM95Zp2NS&rPI}>G^rx@R@og3VC^SBGHHh}jQtT3-0qUiAlDv_h zQT5QhSDS}PVoJE*%T!Jr)hA?(kHC;+FLv9&FQZnYce_1fKcH)1Ij`i3?FQ4n4M`q3 z$w4wDTTH7R#^y`T_=Lbudp(J%Tt$~GuP%-Gh6hXGYRCLVGBR(>1wDJ^1C|jlA-fB) zkh{4I2R??Wde9D>QOP|llxJ3gg1zeMo$!#|a7uPfVg-I;tWYwWLRL&SYR4Bn2z z{{;Pd|Bv<-nqb^md1p`rf6?@8X`t={x^lUDnxj$>_C4S3nrP;;oqr*e4yrP;w^e<- zmjq8x!kqPcH=B=5)80&voT>NzdzMmhsy1z$H!*tSF_`)$eK> zh=$m1se_F27gswHXo7hBGfiHd4tjk5jQ&=yH4r%Um+T z-A5O$YGnAHExg-V?PyPlUhvcm zeRQCiDRh1#Wi!J>lEGBq8&His3r&Ilw+R9U+#RnIap@J3kL~AsOdP!h0kdztfm5K* zZJ%ULcl@)!s=ty^+sQ->C0z#ZR@`e3-5|{nIVv6;GEmw|X8Oq|F&%9zj0;kY;pOxn z-Yna?y0J63Mm{|?PH-Xqgps+vlLbI1g>Oh@gHGwvrrA;av|(h*-BsV~(re9bSNTVm zV%ZYOvpD~~43ed{v)sh5GCdw*{>nksMzVQmvD4z6_d9eCUyKPE3!N%?6G@i|CLcCQ z0S3*_y0%Wq!X4VD>Q^L4)-yE(HaqF8&7eauv-2Lqn^L|^k*A)xG^l_o|58O3zB4!U zW0d9UmT*jj%Q$F1FywqhnCd>6ppH#EG&i|? z{e}LKbYGBe%6|f@=QW3&k}H4G_9s+tPm%~0l+s_jyD2)m8=8ruv$blLUmrBRFme(o zUy=$`^J@W9Fl!|P@n?uKlrZW(Iu!`L6yZF&2_WV3*EQIvynSpn{c!g|PWX~{_@Amp zK5sWhYm=NyKR!6<=B)!izo#j3^uWKHD3e7N0y;>vP-%I~o zr|&E7^W{vMrF4_7vlxd#GpZik9la1tNPqFqGdnyKH`S-({yq=Op>cxUwVG1!sVt2C zyy5XG-*Yonxlzr9aep-61BzI^JBt&Oe6o?eQx8Yl zQmk3d@UzIjyrFt6mK&|4LaVWvAB)MM!|zgA0zON_>)yAHYX1qzeaQsdt0chLT`+~? zi>^J$Gx#-N3@HoLz8W+aXInNXX2YL3hVy-XF11R~j(%X{js)d$dmv(Ph>{KWeDvZI zj3vUZPh(Y07OVO;;HuU@_^;!B2$D*jhYP}8gyB&rmJnDx8d`!6jUiw0`k!NAVDaH> zR^SyevRS77JwvXmDV#w9>-%?G#Mnsu5uha+9q|KEe>SQxLXHS8BVO!3Nwj2qQ|7#4 zQPx-Ya`^_I`I|m#HLq@WwpSoU@0){?0;r=H7(jpCM=VsBO^~_Ec6G9&>+OH-yG&U` zt~oj^I~sbq8pW=Sx-g}0d#~%tK&Rjbu;Zm6&Ec5{%#=*z`7q24{9Gzhtow15*Jr7=eBrhw z1Kro%8AXlxYRmH@N>+ATHB#Fhb?bj2U28FL8eJ7CC((FgKH^e|rK+R3h`-dhTopuK zYU<+97KF#o73c)eB>wIfpqtH0wxU5~0>693L`)5Mv{yDPq0|W&RDt{$%hZthW7OVU z`eoK^+L`3TmedF4Od-DISxfbA71k|xpQ7M!ob6#0T!4%M5HX}W4*Dzv!R>jz5anU2 zknJhIYZxdxI>nQz#k*o)Ps2XUb3`+wM)Mu>A9MgcZ`+3vM)#uVkvywJ?g5=xG8Z=i-3QtI25s->;T+bX0fE)xscZ_2}6t7#O{f*}ITr?_LUkSa|(2=j@ z*6K=M;aAmtEk#IU+lEQ%oIGF5Vrb#A_eoxAsXKBuE8e~N z!I@Q&L7}h3vdyApY^UFzlKs;RLSP=rd^cZVqCOW60(i7}TeF0{LVa%A1J1grR@3bi z$rR&9_W@81gg`q{ntgkR{?BOF;x%Wl62p{s71O{L^Xiw0=3n)tQVO>2W@`0&mM&X8 ziIjbVe--n&5;(${*H{4uFioYt;;8IQ4I4o%I;p5}v3eTvm^Y#xO6Hn$3b$^yT-dQ| zHm$mUPwq+b?6>unGnot>Y+M;bDf+Bh5E{fg46mLFviCcQnaJI@DoI2sRN}^MniHq4 zd=XEntg5W`UA&<}DKJS+6E;+Qqm->0WUiIIgVVI3yux_q!Edxq9q4g~m5q7J_HE1a z9o>?9{x9~asNj3gUoaNWNTk4tgCu2gI1o;MgA4C|q{lq%A8}=4#p}>|T5aS(@-glX zVk`{T(wHpBvZNON<^02ypXX>ABW)5?Qcbxsg zi}TLvXt#JUd$u)u0_(oN+)E#&Zb+IHrHKlRec%sKi)?mZsEQGi(-E$l94f*O8ie=t zwk>p2&!;#^28#BHc^6vvx=(OATb_PjVBLn;q#AKe}XcLTOiM~DU-9N$f}K$!Q&YDe{hoXtf? zd%%AT>8G>}WTkfn5x$Gx<`Aa+c;zl#GMJf_z-s&lF!M;#hD?8o`G!aDpLW4_#Guht zV$ReJQ*}Xw`a_A!ESaVXLUx3~3xBk`w9g=2EO+StLc{<;mP0p* zlDUW;SPIjXwl>^Oid;5{&SI)^a%s4xAys85gHU`~QAK^}v+3{_t#>97lq3w-iK_TR zY<>o=aHKb9APQB@pE`;CW+*4GQ}sLL`8E4cNWi=}r)s1rj2XEIa_PagzYy?MyeEF4 z=zcsIC|VyAI~cU9pDOTPN-{lgtyhjsTpD$O2W`yLe$b~slHGyL%M$!RQx>G)dbD`4 z4&0&h#KTIyA9qGZI*hM3*EA*a#LVfr_ZSst-K8y;3+@u5^=-= zKi=(R2?R8zezR1DSX>@B^3@Ygnm_d^K0(2S;d!D>@jP4l9Vt%BRalK!2~Zb&LS(_B z=_1Y$k!mnjN5@N5Ek$0eQ>L=lB8$@kCQT;48r^6*j4xZ{XY~fgD;?l4d+-V6iAh+j zjO&GS{%F)0j8_*m`n+p7rQq*d+R&c29h2!Dww0Wjtf_EC*4>2uG@=rPGl4%_0$46$#Xl85)r2_7Ueh&QBDtF@Ag{!`WE;bgGFFqy^ z9K4KqtQ?G1&A2|>@Nkvb@Q5&M(Iiu_1vFjNT~MMdIBe3AQJmCsVUhzdC|U3W5p5P{ z6wcx{St;G|pwv7;TEb-A){Vi*#HGzrTm(3Dh7(-CiJKK6O?gi}Yltf~{i1FO-Q~~v zY5BmfFUU~4>84b?&{^MNBR1MDS{gz6$fuG3ReBvWzck=1F|BJ;Mb?UH5ViJbZuSn} zCE4W|f8OTo4iquX47yw(xFfiY5dwkxHKML`FpZ5m(+W=^JFSa)_b#4IaV-;?qX!X(V9*uyU5CZA^60_7RN8Od2@6&B zl+yOdG-WfD%%}0V?*mbdg#60+-!aAijw+`9&!~b%b>2Gh&lSQu%6~{G_0ytg{yRJ? z*1O~8&1IkZ3SN6fnOp0h8WwmZ>3vuG*l`8vhDg^YGCzOsZ~%tcFJ1&n%oC^RebR@u zw-d&{>mS4NWpVygSPE!3o=6_Ufp#wSO40xIcL8)&H*hM=>?x8RWEJ>!_+Q9XSK887x8xP*WcKVW zZ^C+{rq<9;embqs5cq!>7Ni`JZ4(vwYb_q|AAhUOs5#vLgCiD+bec}V7axP(ZM;0v z0)P?yOBxuEZ>^Uj>W%yT7Y;#^6IH8=jXk3o%v!y=&jof`Mvh$K7GsCDw3;Xq2;}(Ih*KbuMi!qDi zXo%({byW9dxVMu(;TbU;+;iTr9~Pk|bLwMoi?+J9#-`Jz413@Gh)R8_AKx9HVTash z85)Am=&u8N+ioaP@;~A|yhjiEG}#z`k>oPha}v||zA>v60ZjfXMJ%}9IAiaYZ=YK` z;<6Rzl~SU#Yai;fm}zZ_>e!gKRfNlWDja;ZI}>HUYOHnjDt(=@76if~g_GTZ5`D%O z)Ho{fp*@?8@PdUm$O}<)qqRo& zQZ?*(iv+YK0hV4DnF@|sV~mI8^EFt6<~oWu>=4_ShnrFJRB);7z$lG)gk_hDB!3XH zGuJqCo4v~DOHiqJS|+d=vYc;Pw|s_qkOn2L^7F_!F*3nlUz<0b zT&DJmVI=gU@=TnU|hDTC4iwpk)5c&%u2gTvN|qh#4GyM5f4uSF3fnTr#JPvb5M z2pC)lj0O}eH{0LUpxb8Ofwzp()xSbq>%YDre??8-k{ldk}B*;tQ@y47+o}!L-3V=_J z1CK#ovsWXFP-DL!>Of%g^hljlPZQU+*y3nQtfG;ON+bXh z7yp?W$W*?YCgIrkapbKZbHVSfZxB83#iM=6fy6lcA;GT^@6kuXf7pGLllL`lkzAj! zlSso|o-WR6_c+FswIEXKvnhq9It$}wxljafsYs|pgktTi``r6GZw~#cw&L|`9UXl+ zzV@;C=$94Hjs~&Fq*wfwhoE#(IBWP4ki!Ey1SU>oHKI56*+wP5;ncpAVM9|)agKRe z-^Zi&kuy?m7or+2r?q98{z4F?QTu`T-ObW|G%t}KyOx0NYkk@ozxb-nl3sa){=en! zQ>Wiu{bPUs;xoG8a{6F&D5Vy@{<;y?_q>r3UfX>_$=%_^7tciNm%{A1>u3) z56q%pvzO{7Kl?=YkrD^)RK7BGNK6&alv>^Hwt9ZL;VGP?A)PpCZ@>eU3)Xg5xi(5F zbmSa{40TK0rTbOM`-V5|=CftnZD0x~Oo$;C0pocw-0C}A=`A6#qG~Sy#?Rk(HT5uY zi78ovy%w|$rfGg9Lq{j5e)5D7GRTYLC-bPT^Y^Z$t}*V(cb*~ZHxm7Q3%rzW$WAtD>V>Qw`ib<^QeZnzh3EWPAQ3iZwjx# zte@RzntWIvE!$T9@wKB-{J;>%3pGtxAWXVLJYSKm};Bl$&KicX7vE4Wo}#k84C z<_!h1DkCI#H}0na(Vz;|qZM&^G5NSjsaqtQ-hr7*Yo>(#9rSH~q}75=FHJh~JS7(H zH@+A*h-zSpZf60OwWe5kwm$1Ik6_dy<5jsjzH?#QIh;7qj@>cc2sY0!-l% z_`%+M&N-??gjEgE@r0p&tf6wY^sBIWM%af#PJ8L+Z;aDROoC`*hx3_1c1XBDNkr0z znG>FW(e>R(Tjah%eD;ADBGb_{sTN3k{eXLTb<1(|nwgYzbb)sF5G0z0v{k0051O*C7x_<X9cFnf7iZ0Y zn2gH@;MbrET9ezDs%i)7{m%3Mnt$MdhC>s?Bx_&|im|_77YG5fLYYe=;Sc&hfAer_ zOK`ZV&3@H0Ij8`y5BrG5S*!t*-cC66GbgbP`}D+XaVEArv{V>#-sLeiMk2(M77~LA zyH)#Pi{{n8-F>=@gi)O0#qaE0#2I(PY4sqU&H^}SQXzgex7^XPT%*Rvl_Wn}cUoA0 zPa|EmFO)75I;=N9f7gok?!+_8!pJj69iR0Y{=5KD*_cv5Jx?7+Gm^E-$Y*D= zuTF>TJ33fAlCbwMagF)f=Y03U0JcEcT}DJfTm9rQLJ+yQR;!NQN8vUta5ssGmh5W} zj(1VsLE`+%+%<&{Gvz0080^w*2AJu$0`+0fV0HgyeuTOi0VoQp3(){eM#3U=@Ne@){>5;i z#o5Kx$yK_xkosWVOtqRe)vWZ?0suS=fnOsFfLN)m*o#OK<0ZybldtPEl*arP_Pyq? z_#NS%>FKj&Ux?5?IBSDYwtd8z*#apSPrX7|rzBEoh=YXKS#l^AEKWn>?lAY}QMOPc2IF4n^+ayUA{cy-i7dLtMjf+M3FGOC3?Hso%*N)#8%(-rv zpE(Fne7PrHVWrMdId@lA#V!UJ73h<{(VfUqa&?2NBdm#Z>#%s_g`?sb@|7a{)s@#T zMv}+Wygqkp)8z+fi8fb+3#)zaZNJI=BJa_PFEe3)k9Ux+7qVW&%InaYQ+?t4mibyi+^ zXVJ+^$x!!BR~Uwj_4cM zQWU%^7^G6X6$p;uy6Rp<;sx}-^JU$#Z;hGAzX3~tZwS@2Orq*{!o(9zw_hDdJ(q1C zT9^3f>QyEv7s`DUMao6wAexxS+@Gfio%rifo8B-%OiYM{(9riIsVo=5s-5oaE3>jp zKI7Yu=(Fw`SkdlFN9uv1J~{}w1h0?lw_LRD-N5uid7BdIE&K*>txg84x?D?Fr;MfK zS&eRGw%Li4)8o^biMv=Bndcc{uob@(e|C~8ZeC=cyBibxd2)8zz@N*+#x-p;VHx0I z?cbzXgXnM%?gk^Cv!(@zGizCy%~ZG|3~76 zlia^|N>3gfASEn6+!+k!HuVhk_QE-y-FUsl!qqqnJHI&mcs-h?dLfM-2_79sy9bG5 z5Xaq(;WBEwSMkHY+5l@Vu#rD`XG=nbkM|KH!|6SDZa;RGRqY!;{o{bDUnn0^eCn{> z0B`5$NthaxnsE58cWqKaN!r@DCBrJC%O#m+$Fty+E0xs}jM*k(ET{(_M0ea&zr{wL zu-R1Q{qw%`r+%70>~yUkSMz_%&cGJQ>&d1ip1@O=dNrFg0v1=6 z#$6O8T>UhRCU~>BBCD=2$`1{Ri*Uf_K~KHGlcX4y)Q^;&C40yuDzuvxjT8*;uGm9$-pQ%-%A%>TMOPKtH_?qvMa{GyIg8k}-4kTh{$6WRM(#v#{ zV+6ZIV`WbO0ZUIdY;eNwW5X&}!=kmVGI={}h`64sy-bxLfm{=Ba;}E;+eRi7eZppVI zR!lWDkS(Z-{&wEcvDnPcD)9QlO<^mapH%n4dA@g^tFKEpR09LdP*d*z*iVo8ESWlT zp3>Ux;AFRA+JgQ|Tk=5KqUQ;Rs--=5SR+dwVjrt(D;yLhYJ14K$!rCJoX0n|yxYZ*1+sD3Kg8>tc1D|hS5x8}?#Kf9Mgo{O- zl~_mkK#9e8jJa+eCbye^q)s}+a86WAnML`Er%++9{7mt2+~!Kt>O*c?%^4>~3VItK0g}iuA9Q+G$ z&5lR?+Pq{>d8xnzAdNOpk;(n`)}oh}tJe9hs`5r~(xPQ_v1( zP3naa<(>hr%mOGjn=x64I*Zvz#JHhPq?|$4((MK#TflJrh+0bumm)iAy%tJ68wpQz z6jJgg0Zp#h3zSUMw~R1Xkg%nr@M@@DUgRhEFOznV7DNxhi2xmZKHxz&FeCV%3j&m^ z;YJ#n&j3*E<}&_ANB${w1gTAQs{WRv8TAmu^26L$j*ye}%3HLfzoYNQ3t@MS1a3G| zL+jqat&Nv-e<0P%FrtJ<@^512xd?q?#Mza?m2VS2QUpKW-d@C>aR;7-^&BK%(u%b@am0x@8Z=o63cX%ZE(pUaF9?fV9@u#IiiM zOc@sC7uR46TuOaY0QFk8?OMa#tT|7Kqe}i^HyHSbS>`RGhW^Pi>Hpv*3*k$i7jjZeHY6}J$WQp z@~43b&Ulx~)3ly(%A4|U7IL9k$7tUN8&OwlrS27`Un{hjpe@kg{jV(&{))ihy3Bv6 zZr=51?~=|_*AH(e!>%K^pMn|efr4q&7R8pV>A|>l%NSp8 z_Zvh48YrWIIqy2650j#afJSfweiIQ<>8-(4Z@Xx2sFm>W<$-yMt-kUW&Q>m4nLEISP|)Co1yGptS$;dZZvq(t8vj_anLqdP84 zlofWE@vY~8SMrvEBhO=OxQ{ zyp)Is^{Lz7ZsJkO6)M|v?8#TP>gQ$zgQZW88!|KN?{!*#YfS#o|D^0b1jiyWx69vi z$V*nG09_^z^eN(IzZiEksdrVzDDPLL^|^yg*(B4gs6k1NM!O}~$l_`#1v8*|3MVP5 z)lerX@J{kmxQOB;60mT*+860PIX&ff#8-P!xT4n6G<*3^ySs23G>_)pIF zVBhCV;)A5~mv2Nz7ne&)zU5L~#9TsoL5UZ>Hi0n}KHG_CSP-G(e;N~a?S_-nco`*G za4W@a3GDC%{g2HqT`w}~AuJF|g6MhgpL7HO&5wQ=7QBh>g(d*#V#8kuqf^)uL<|A8 z82YI3`23S5z7&zF!t;W6C9d^!Rh*-{a79d?xJiYf$}jRKt^c~LJe?O%KgZPAjsU8FodTldf>r&5cs;Fi1H1`#J{_u{og0}C!E+&u%yC^Q* zBp~tr`y2&C$zFKfY6|YBEuJTo?{;fvCT>JU+fr>{eW`1NTh5KoR*I8xnL84?rgVxS zLK1vGnY9?W#>$<+mrs@xQ`X==%7qemM9QaW5?|E*sSX z#Ua|N)MVD$+P9oWK)IRznnvD79Wip?iS4lk4lxR6YG>O%`U{!YxxbezLdnv_J9~PP zgs8(gPnR`k;jFrv?1Y0%NrhZSn=i4}Dsk^1bCX@Jsm<8PGjGfXthiJM*FoHaSE87l z6$J5*8l940O^#b{Gf`p7M?sdenD#7jx_pCu5Itx25u=CwmPR5;^8_D|~NGwVIVU*UNXinF_ zCYtlni}=Csx$;16y%N2@E0oumP`u`C8SiC)JKN>iKwR! zUL0N{CsO+c&B#znGs>4wG%a&LjLs0C^pK&Gxr3m@#fr z3OhN;p}6BzUBz{1j039Ipm@ap(*FB=*R08-i+y1e^YMk!gD;1NC(`--SU%DHctuIV zxV7-3pCg6n2|_-Qz&(E0D`3&ieI8b8+hb)+9?1!1n)>8$lG}Kx;^#48p3?9OHlKYJ zmzC$b!HpGOJ56~fFE5$r<14@aF-vKyIrW;zaAH|`vfI{Ox}T+cy;6n6T}y~g&Y>CU;v4M~-AnP;O#&e-_;=VCB847kR)7lbAkxR+&8@qsNkaGUN3 zO6szGS2B>w&tuZp60rsBuG8{nWGQ0&!e7YimF7W88}uhd_Sf3Lzj(B}34(-ESj7P5 zl$yN1UAy>LXvjVG7eo52Y%Pv8tOc}v=@7a|h9|~Ronm3a8(C4Xm`>4e8qUI#lA$jh z=Jm61hcB{%Y9iT0*k62FeNzgqWF zR-56qpg`thnMw#U5=0)6P%tk%yOt##iPw1dY>Afb#^6*ASIp^dhrv+G2KNTg@u-L9T8aCI6m+PW)b>+cbZ%Dg| z7kg7mOxl3YKOA|pnTQSdMU3a1%Zsj)^7rOyaojo|!t~zQ0YcfYG9LUI7ZqW2YduBd zkY0o{A`3L0G*CV$;`#S^Bsua;%jYMSQRse60wKf6*+TtWO(MJ=G9hIb8}t{VV9v!u z7pXetg#;8@YbRMkAzmg?SNUVuSp=(7Sg1+1{q01&e^Jij%FJ2oYXQ9Jf}915S2z|f z=+o2vx#6KA0>sUbkR4n<$`Oq-cRfRiZ-+?9#LlG2l#(s{27%5~+LPmb5=ridnD{g> z8q9204Hxa7p8R;s-7Rty7j!jKV!pQGKW%q*eW^&3ZA<}t;R-hGIZ_olo_`8|9QBZ9 z!qeoCC z+bag4Jp47@3L^CuqTEH>0#7Abu}_n-r4lA{M=6wTaR?v zXY(^i;nb1FoXWhdwOZqt4U_vRBfYweS6RC`QVF&r06h{vJh zA`CajnLVjzr>ZJlF#kE-o;<%5r6_5tq{+Bww$IM9p3Z1O8wr1dAxcZS5~gwQnCSL& zssLJZbND)VewXS}e^Wf4WjJ31H+r3^+cHL|BMtXBA* zo7E$5@LV>sG;r?I`M=ycXZ~NsZ5-vJJ}Cc6i2Va_Y%Ly2ed|$u2P}gC)TiADFsQFy zs-{Tfx2OG2Omz5P|~F7>9??*HIJS=l+Ett1!fVHLwS+Vq8A2@v0rDrY*0a@HteCvI6L#e=kxR zcbr^*QIVTyY27sQE6+$-zG0;}NsepaiHc67i>VJyB=@{AC z+CoWpNqy{Wvah~>{oKs!^!Tfm4F;blT7Z-`m+21+WHxZsFnsDwZ9jnjx@pCPGe^|H za996A{%9dUi8)^>L(O)W+q{GvMRjxkg=}p^_=9|}9)05Etq+@$J|iAlQvC_&aHjRI zCBRsbY(yJ95n}XlZ_sXQuR%y~>9FsCkXC^wpZXhR3ED31n(P2!(+g(l(AkLXe*#=+ zf)Efg4JKI;*9MWaxNECtzC;+S@!y`iU3b~`L9|w)$`HP`wN~Vx0-jVbLLE5_Z?N4L zgL^8*W(mdhVNQEB3RTCejp-eLVdKq~j;5T&qpxRPIS9vPK!*pJgkP?g)4QRZy&}4F zqN3)W0mUb9(@hT-(l#x-&gn_)BKRNDZHyC5Qu7#!xkiLPgWvXZG?O>w^#zzNgLJJ~ z69jUHa8a@(47P8u$${E9{Kx1U$meO%aQK{5u(rH#o!7BnH=xFtb-nyT0FX+2xQ`gJ zC4+&`WvOL7g|ND>+5)qO`dZ)9r6Fl9%KQ4TrBq{HK{wY(gTWg_+n+_GhsY^}SUo_O z5`1Us{e^z$TmrI0YW&l`z4LTjT`#RnUh!^NG=G}RuhMd^7z~_6*uyOM=t!2-FFJ?& zp6iHHVYyhbj>_*sTTL3-kDjC!UD|`f7D_S#MSD$npT`%gY(&!2MH@iMN#gc*F}<%SFVm!K%0TH|Hqv{1`qe8(*>YzAP8I=~UGy_p4Sb)a z82qnZr*ees+(Y^)U8haa%abIFDUUq*s+@8IT$)!WX=VWWRY&}Yc6?D0+VNQqW{9qzWf8x3T7gdkKOP_GeX z6!%NQA9LhJsJ1uqb4elH=J*nX4)3b$eAc>=Zn`J?3kVF!l6a{TFfPzHi50^1`k#Vw z_ZAS11~0esSr>#=?KvB{Ho6KbH^1}|_juNm`1NdtCPaE zIb#mt$do=!^3Lmp$kSCbGG63<{pBE<32=Kf;V$UKD;w~A84tW978N7h>mZg6zPWqv_pzDwUxKXZnNkz+m>=+m z-@#6kop`dAyi|3jgeD3eJe{*aga71eDGe{h&ekRhFS`N}JtOe5}f z)MfG{6g?hI{&3r9IWoqNg+3}08m|brqkkcdd9RiUXk4XVDW?hkPll6&n*V>B7a|UB zZHjMZAST!L6mggV7$e5rpC&>Z|Fw2uchR?Fw8={HPrhk;Th4da`y&3gUR{Yw7vN;0 z$)Q?T!@d0%W2q-kjr^1JE|?y;KYn}Vqw^K99qD*GY=*1hrwyCSsgV%b-)x(u`>Ocu zH7ZLJ&T=42Om+GpkbXelXsMkJk+mP|_}wS;a89|4qc7Z@iSsPsHv1M5*LYnYUig9P zf$qW z@CN_=bD)3T8!kJZkS{SK*8u1A%;LY@bdvXvzB<>WyfmD`(U&GDEfOOgRLT&buf7X3 z#i5H}r+>kC!$gQTdl8(!_qmO}lr%Qh*H48ACkdS9Xb(cKEU9^oIxQ)5)df8{Q#MFgY8L5L!w(;#{ zH17J^l92>2l@GDgC{r&9N7ReVf$q(JgZrBJO>0UX%G2o%0^e9ti6F`vSll#ep1x0d zNN7^&ekhn}j#f^jZmXP5&^I_^AcY=zPx$~EPz|pd$#!^WN_Ujd3hw(C5%cIvtyWCC zoV{ITS_ zGXl$!jMXkW0*SU%EGX@8xtW=p>H?696Dz6lxy?1x0!uuYckhd87QPI!!xGh=_Ud= z#d#^nHXQ{p$Hp*q9l+C_jr*fXX%?_e34nSGI zZ8Fll9H<5uFU%r_kO2Z(3NgFXF5(*g)GOB#V%*LD8tSJSF6PQU@cmfj7R_sNH$#%% z=vlDuh=w`Wb?<$}NLc;O;CIyeTZC|gNBE|Nu@xQ~^P5f9LiJ|r>+x`34l<$U8sB40 z6BB#8fzkeSbb3k=1U8}Y^|R`j1}4D?PH3c*{UR-eFZn0s%c>Yi=zPo>+p5WY&MBlV z?dCr8)!rUkQZ&y6$w zI~e`h3BjVU{$6-icXnp;5RkQ$-8hU#G|cV8KZfl%%@h+$u;%Ces;dP0-9IcZ8VDsY z-%1q_)0()&4*9^y1pKs$!5A9T4U0$7!0$J=OXGFAA69x>&G`N1_Uv#Eb9v>Oa{Kc0 zGS{zJJvCjDT0T)HOo+EIoH;5{PnTFnQaoFS=fBf}XRLVd&EFd2|C|sy5J|@_Lpfe)P=^-Y`z5C>W`3UV@nO$ zT-%f$E=}3TMHd3o15~v8QD$rZW%sP{lldsG+SzcsgM?4+Rr7`+>V3>u%X6zM<`2Pm zlH*5hOh9pD*Lubn>O#|6!`i-aIbzw%7RJ|ZUAt^{^k0%~TcXlubIzJUiNbR-F*d9~ zYz6{a-kZi?Pagp#y9%7^z@tv}X}NY|tcAnDz+cF%#<$&1)f@BFC-pf-!(M94Oa2s4 z@>GiGS;`fMP26=C=L~`dBF13*UrFrLX&5KU%U=+5+PcKpa{F}n zl_ld{bH=8&vAAf<&7ZWu6t{ZD3UgB*mIB9_SuTWbBD!pe3DpsfI8vC8_u2wsHtT3a zcyxPi+*3+cSmin8M~rlol$a7D%KfIF)(PMqPp&G?Mf`JGG{RTM9Ow;xeDH*^=&^?e z&Cz!B$_k8~HLutM*pa}-NH)l+?VE~Z`D#983yZ*v39nZyo6CgoNutc#fRK;{nXaWWXRQ*;Oi?rg{~<;L*P-4tj?>xU~V~ zg*MH_ftq)A%fM9FMrBOug42*bL~cPBa)}{)9|&+GB#Q>sIQ))D5?FJ( zz^9Mwo!)j?z3O2<^T$?(PP+j5?TaaQW?5H$FMylee4cJRi+y5|{ ze61zF=DrCrJD4mg2TVTp5nzAWEh)YQ3^+af-v)LO3ZPvhit{fB&p{T_aoEH3y=+@{MyB!I587I{FGb4GVycT%xJwaFz(g${G?o1KoJd9ywv@uJ#8)^>Pf%E{ z(Tv)z*0V(m&M@f)g#6^yj4#n)?y~6yyXpGV;3|H|0RT7XSrZF821=UkDiMnJD#Bb2 zuQrn0%G|QHkL;Z6Z*!lOG&sFFYNSFNt3CLO#c@2T+D{F`x-0Z5_i-p+Rw(z?F1ol> zjSIjxhyE0EKLjPA6ZcOkYO1EQr3Mk(WGlfUrYvWxktY0H%Fjq8xqQH&YiNHFQ+znD z3Htw37UVy`VhH}!f?UY{B}D}V--*6_<0b#IEp{t$!ZU6(I1_eeVEjeBeF$7vFpF{4 zoh{Q$J;|;$_FoLi82Axikk6}cBLHGZ^%N!<0gkL(OXAtM?rP!htEwP| zT*1n_>|F1hk>62X8}~1@c$6O7>}V$Nhoz&J@o8z`#Z-_yZ7ek?1S1B=_eMy!sgwJp zD-k;N7~(tZC9{zmbM?Ejx!x<**rP@7@QUZLx57LgO9Bo)$cEU9Ihlcg*E#=vWs*Ty zIH3`bc}`ZvvpIfwn8&a$?<=pi^ylL{yT#e_3Fqkw>-aMg?lZ)TdjOUq<%YGJ?avYt z6LJ~``<)hrs7DrIVW2|$diX3t11>5aGb&Onv&a%akU#=V9uJ$6g0X-RlIF{F_7gCcY{y7*(adn-7Fh%W$}!j&7JCb8A3kgUBmcdZ&$tuUb8O+zey)-1V0lUX^mz7aOE7kGH{s`hs{Jkr2E49$D1%iL|`$==-R_Hst8exzI86d z+{AfOP(b|EzJg1Oa;zjxp38+d;(p>>MN%kDOheQD4U*v^bqdB^KcU>_aoUW3qTyLyxejs6($Y@9}!QQB;JfbYM-PsKCNiTR>6A-2N3rQ1`>%>S7Ss%}J}98@*yy{ATWac|TvpV6rTHoF zL40n!po_Z}UHPTBNlXb%kD+95^ZhH9k*({a&^CW)iim;pVsKBIm9fpR1mT$ zSiQ+LMX05eRVut|O~M`~@nQkDsdDv`*sW+yqsdq9>+WXh#`LHAQTNx_LMko$FzoGT zu$Mgha$?Ji=AD+Ct)kxNuSliuJJMuHtq16^LJrH@UWWWXwY_;bl<(g+K9=m+*DPZh zDO<7=24&BZk}ak}_I(dCmdL(_LLnrIN%kdUUlSohmO z|3%houk2P@%HiPT00*1&80r3w6(F{M)(+ro`a{}xJb9@SU5Y7tFv;mJl7=i>3@x@;u#0jYt zFdi;$Qk@E+e+-ftK|@m zi04TKsD4*a!=-Ris$B*-B^F~E3Hlkn|K>TjC$ca1UXb-0u$m z-8KURg~rDSu)TPq!FXyaPsgh&jMF)J z39Kid5CE7$okUQ;HLn@~UIIKt>3IO+>{9P*`}vO|u@Ws!_h91l;>@}&XQQ=YkbqX& zos497Gwp5-&Z$3Ni;9r4cIu#lcz?|^#c=QE&w1805ohIhBh`fUGS)6uLr5&L9=!(A zZgpJ4-9eA`x{8Gn-HYvc!=oyP&6&*T(X5do0g(Ce8)Qz@_ZD+vTr(D|v6U&2{42g3 z{i~+zC-bH%w@hxzG)}WE4&1t&9+ZBINuh4UOWcq;_by!HF_BFH3q=rY)b}<1fmAnv z(S)Go!Pr1;mZ}Jt*Nq*qlUAQ!8YQ#w8y0&?$KYc}c-bk2sM*!HNRZ#aUPlC2lz_=p z_o*272K)FCS&4A4kNuMod)NDtZr<6We1-XPe^??8X4y~}*)nR(7LpT-yeqYV`C7ID zR-hx}zs^*iSj2$+J};r&EIqtoQT=enLGzqLp>1wO^CROO{>yAMu+%2HtEm`*?3uI$ z4=BA7tT}T*PpF<_g!Ac)fpTiHJ-?2KO|+Te&2);vjeH^pa)f+(I7xUgbNl@wV?_J2 z^9P@cM-%O?S1h<4TLs9QK~|&TbvGtE9Aq#_*Tgau-Hnx(nhhLqcca4V_mb(Va6U`he8DV&JI$L~)s~&jWW(m3Q}&6`@6p?&ui2c*AvgFH z&OC`{GNP_?gx}x!3)RL>*&y`r-~#!8FORA>!ahnne<9?`An2r2uit&bTpk@jCHpWL za*{!&|Alis7$fxMx_rOS-^zrEl}s7y2ZWGvM8@K z?uam0o|MC-Tj35Cxk=a0P^k$(Yye$z7!sg|3?t0X9Q5{|EXI@C=gobi`~@;cz@Ggr zXv@DFx*H~(ntDGP-m;%i%n?bCM|B@MY7#u(cGqs&CG1Q75bSa>V|cG>tuOdxT5f1) zK2sl=av}wcfq{{A!+P^Wm0K>@lv(8ax8whYb<#+3%e(gs=C7c6V zN!^-AdSt(yQ2ScScnp-?hb-zbpPox9jN?C_Seydh+=W0`1QBF&Z@H_TwjOB1ct(gB{eWF$u(IeahDf0Z%9HEUkAX}&dG z-ssli_sBjbNN z+|hqBTI(U{8ToI=O;wD5UnyKZL-d>0FkR(8tIuDfV%E`+*TY}95CChwshC$Mdo8)* zW|(Z3Fs-WQ#$FZfDw+Q1Epoz++=xf4i66yjAgTlx>(-^za2d|?;h66NO)_D++?J}5A$3w%wyKqy^Vj&;;% z0}8h_oBZpu6;D1q|MHsl6z*{aF51 zNoyd{lpv>_>bjlV?^#aG?^MM5Jk1$@wbk4->Av}kZiMG{Hr=DQS@mqJUN7vaFLM2z z-QA7t5gCuH-Kh!Kx8 zZ-hqnv4_30iKpxjub>-Q*?yoO^@RwQ?N~kHIYI?qst>^~hJR3BI=X5<*GRu7Z0M3C z(c&wAKTTMcYRDF61Sw*=0Eu|qJ5G?mbT3mNKP}0UiV{kzw$)wJDmb>!J<686b3E~0 zv%atTr4nm{n{KX1qR9L5nkce5g-MJyH2DyP(}TuZIO=H5W+urs1Uy6fvNaRWre$w$y8aRk?(r3S(BGRCw20J8~ogq zmOTD2cH=Zmord)z9OO@F{^;WULaUbvZb1hPFS^Bjnc~Rs^oX5z6bUSX9W>k}Y zQlO_VVDP&a6n@-{&x5%!&;-zMa^Iy;rOu>-9ut+DI)JgO=#Vq4hlo}(N1^SMq=zT#v3{omLInLHfUL$h?NtBL(d!VnrWzYA2S zR|x&52;|r@L}>@b7-@@d?0o#@PVYudN$iiGE-9+kS>E@b^zo7wPeUQb>r{^;FOgnh zTrbVv??p^>@`Vj7^wGO`^xk;v(*LB&TRzv7F;GCBv%5Qs!U{&IBL@F8OXTVtI9WLC znf@sY%tK^HSy>P?21jPh8%!YIws zgV{&@J--ha4lN7o&#Pxvx~;5GIz$|XgAuTw^L&_Fglf252Wf82Q%cs9k9eY8tN-P> zbh0sxYL?M$Dq$daPmFXV$egm9oH7yNICPShm}xmt<#T2FhAy7kY?eob$67|rJS8}7 zUO#jAVpyK=eX-0^Uh7{x(w6;F{qj$;=)qa?R$MIIAoar#^q1wd6a9L?o~5oOP)O$~ zT+7;-FMqGl*Cl%A4nmRHNI*JS=j!Je-VzoUQXvuo3?5bp`zxbp+P{_79BsSJc4uoo zc(JH0QiZ2!<73SA1N&{XD+55Fu4_qt zwi#sS3CME1_VcTS8P}~mO$haiGbo>G<=0dtCCu$fU|4^D> zi#2Bq+vxjDaP6HB?yz%w+N>44_^ryDSHOTez~uY{%@G|JSZ%p20y=rf3z+wV0|*}b zFc$1!#$Kc8L%hREqo-&4?fg-`d;E90*%c^;EKFGGsUT|y%Vfql*KiTS)rXmeGj%BM zfk|p{ZX>O}8%CRaxi-<^B4+9W{|e+&-}UuHd5tHa?kCvV@z1#gvgg#X&s*ME@hdBp z$Z!NRRsc;)dPS51Z~HPuq&v1a^P5v%_9+(X|Y9~IrOnnBDlWj;f zJHf0eby`eNau2K|3geXGe_!(U_b%Pa8s$vB@p*DaZfHdRRxE(!VlQh$ImjR3I8&BK z|3K=Yeer~f^01WKc;-59Uzrcpi^n&vvb?B(1=L*tCK@Fi4bDXzv2Y^;V;7B(5@y;} z^S*Hkq6jYPC<6)Gk=sr~0DPSL@MvU}RcwtU2FZ^(@4w?_(KXB{-XV9yj-1aG-x6lmAr#mnB;vJu5vczWYX9R4;m> z;UD}bLJr?9nwvL$@sbT{pO&|GyoMVo$$FAy{Ej+6o_-y9zrng>Lvb5IwEc^UgGj;_ z%!}2*L8NeFW43`C_M)fAHEH8lzj!}=j>%RV@OMO?8j?q$?_GPB=V*H{G@a6fk^OEm zH!f{^xYDD4K4SRNhP#l{VNm*2l>(RCXY3mK9ck6CFgJJbg=EH??&0vROK0xb_w1j+ zB+h;cw!AGjlXFJafPsF+6~jsKi*^@};~~@6d*Y|M^N}2co!4O*_V1@lTHZ~ez0XgT zw6w=`46)V18j}KCtmb37LwcC9DNff8Q~uU2qByCPEVW(Cy-&RH#!Wr10d%YR?{U8C zm?%j!?7W}ds;7NRAhK1^6s1yJZ~CEsV^0v*CZn}v1WwA2iVdQ&E{4qKXSEA36-mY# z?NVr-LHS8VmoPsCtj0(R92J><85g&5DW)TQCBnhP5>Jg_^GYwQ8VeNt%iKL!&lpYA z<-ADRVIPEL)9NH2DmT~4$ z3V?P+uEW)@$(F8N;@%onJP9cb_f&f1E9X!VXW~hhmCKiysBC_R^;JHo5l)eJ30F~q z4H@qIo^=ga`gApMZas^Z6@|B6x-SfGi$(h%zX28a=liw~qXVb6gpv;}S}tAe zgPxMH&eC?8$NJ)az+K-QBeJMbErGF*8tAJ+e%$obEYGWOT#)PV z=S{s{JygRcUpKhVSv|Y5ycd7wPvxFHK7@bBr8&3b=9?jQy$cHf6J5K@B5uno|03Q{ zELLh4I)^s+vH@m*FLH_p1QqwynS_?{mfJYw_BZxf-r>4?IWhBR4m3F2ivd!2{jY-& zbzA!v>Al&*Q*i*5v{@lxY^>R`=U7m!Uxn*lK;v{ zdg(UbR2YsE-2Z*NM;WO#MPzjeeU^sfSx30Q1G*u+rZD6 zg_^A!MD(>EGU8~v%pWg9_IBZ}zU<8@9con`JSxrVPtqgj(9fKbbztHaUBd!5E+oFp zEV)eGw02m7)CcDzz6O$h@1gaq#D~VbCH{XP+=y|n1O1`;KIymjflV*dOT$I^4xihC zw_;Ad|I4cVUv};P%>M5}YtTIx0IiwgGyVs(_V5qbjsLGgYrwvpv;L5eka22(ARq>p z8GO87efiDkE)C+)TCY*s2$^4YPS91}xBSRwlUmDAl=hZ_*zy?x3JGp2V?l)+Kc?AP z(D<6{4R+hi`J{T(rri|yKVNsUBE@mB9`Y%w_Z8f0|AZSbb%^|SDEuw(1=Qa!1pWI9 zAY)?|K@e*u7Jy#K*ZwNhW{cGC*?!c4v?t=vcYh$on(w?vN`$)#A`a~dw6BOtc;+(h zinpV|{;v@qI!)Wm0~^9E@~j@*x^<&9TKv5>BwFhQK=+bQ{sY~6^}mhoU3eMIKt{p6 ziwPGJYH%Be@pzWrC>1Q?QHKFbeBW)j&+Kl&$+o41*@7(y578ctzb+uqg!E$Bq~;Qn z8be(HM3ZJBN8n)&q3%?l(7fu~TPSrFaTU+sAI#=Cv{#Yz$b{qCmI70z(>K0=3Y#F| zXo}yc-3V}8p=`mAcVSY6E>Dw%+|@ak8Kw1Wy%o7mxt_0@3sl@3_5Kc{`L-YVPEsEt zM!Px-UW^&?L`8Q-H+%D_CKYwK*dp5dHVAu~Ud9&#}75E0tanV8S-a#AdvfX@|x` zP}eccfA-HJConWb=o$#637N6K#OdBt@P6y4N^o0yRxW``w%zuLnnSlvx2C$aJ2)^u zenE+QzT=_AgIg*rbLe<*KJU{cFYIzQz3%7g?DFcOpdq2yR$svFW{dL)9T#L2b0d51=$dWgu|Ym{ceW@-kr*{F2xXo>G(x z0b$iXSr&D5@8&n-8B0YL5w{0PkWhUWihycuSf&w)g9Lt!CyI?c3upRM%t>gz+VDBp zf$D~`Jn7N!p?k0X`!}LPA2)U1Su!dcexv*J`B>r<@EN;hFdhi*$WZ)#Y9z1|Ho|zPnH&b;G4V*jEiK4Iq=Wcj z(MfsE4vxR2jNKR~?Kki?-u^QCuFX0tWTk)2?5ZlJg3h+T&aS6}_x_VS*#cZIyPeo_;0U9<{dH8>_=lu&6XfsCH#p4h>J4mqVl zz8}fIdS_}LV0R~mEQ3RHKHk2w*KJu@N1O50+a7Dsdlq@(aMK=_kcEm+`GgST87G@Md=!{1`5TI&Qt#O=!uV$s=s!WJgn zK4B25qKu_Xun4G_PB!du_neh7Xh;|s`J$^w&Gz<&ep$(_J7+*!XrjCFv$nwF#E!Eo z<2**xFQg``eT6%U>yW_m{)p=-5YJ2F8fKB4vmf*k zc-S4&fgX!OV2jVH1C`sh09P-Cyx}X}f@t%Uv6j%KD@;npoH>G*~)W)ut zU}7{^56;E~?0G4hEmr;(!l=SmD&X&r`Tz?Es~lHKlvSSE5fb`xW`9oBnKc~WVaiCI z`>uMlMVTLxy0Q+@hARUU7Ce-M#+t}roNN3a#G@BXQ)PJWO7HT*G4B3d^j&XV9iR8G zw?CENs7kj7jOJ2vWNpGhATKBUHZU^23M|NKpX1OyNYTeVvHX`K(6%4`Ouzk{ROm#L z3ez5rHw!=WKGYk*M|utn6rqh0@nm%ZOFvm<>Ng3hi#x1$sdg}Hp7^F}!k4!qI>;t% zr&~xdKoj71=Ap`76?pc`ARZZ5Y9nO{&u}xN<}-?L_6DBI^-Fr5Fs96M8IJNkcAX4A zmyyigGRhLFJv8AwdM_~W)Yy%zh}+wU4MNe5V#B><;j4OVP4&-w`%*qq*Re}!2P`DI zNXjYyI!Qu2f#a_Ge%)W)-);9^8$+zl&&@f6J$ozM)X@0cM?k!|XiTcj1>z)UWf6Fg zt<5{Wy_zfREu7&2A@+eM^aME@il~>3Um0UD$34he-9dAihN_%r6Sod)My0Od)675tYrGU2@+JNrC%p3F|IFRCYHf|>p3C?*x4eW<%%byWnC5c{(2L=?Me)Lqw z4XavGkl1F&F^xO}pqJ#D5xpc$=ys_{d$6OO02TdGJZkV_d-aOj=(~;k4W&<+*h`{y zgfmJW8y{yJ_7OB!%`=MYL(@7r!}P26OTMt?hTCR~Yx2{vpryg60s-Pmc5##G*tFc8 z&4(^;E!E##KPh6$iuJRpX79UX36T(_*g_zYOMf+^=?K0*ED7`Io`Jd<>z+=&1{`VH z(yS?fwqCl&S^oLkY+7Mi9ZR(#H3&0!!r+IFmianWQ9ZAxk^EtvJ-v2yFPz(Pq%xb9 z2u1n5=eFy=IBT^*^_doe`+4HxK;r^T+&{q2lo~mqjNw9FwEw!L?wIQo%C5&X9cuZ_ zB$zKzGBJsPlfD1xA%rU1`Y4Vj`hv%~sPq~at-@-1y>Qon1hsq5sI!Jr<*agJn9BWh zsL!n_F2_R|8|*iOHKE<}nrYwLX_DQfWmdBUf+R|;S3O>F1YJh@>28i z8Kja#PmKb$Ie}&&6iMazCwgz zY+fZ0GHBx-&9vS7vRd4G_+#?oZTgqZ&IQV@(sR#^-`#~!JC&lPz;8)Fb!#Ev#4rL$ zq8}cAd&aKwS+$K_FD&5Opr=ej(`&Kb@Q+k)q8V~E0_b?Dk5D*M=vUV@)b1SqD&Vwh z$=J_rO)RO@1mw$){Q9PI#n4XF%8D9twwU-9V5>Z9HRV{zAwiGx!3x@ z++DIw)#n!11M@=eKgt6s%Gu`UJ3)6ZHWv9gTC>r z){thueSE8S@2eLjMfWEP2yuc`gj#u3@Tm^p-tEXct;I@EFln6qW8y7S8{|pzzB78c zrP)?2v;8&09c~AVd@OL+L>aDCpyZ3|F+}#+!z$c3;Ez#)eE9j0=Px%qp)+2=SE|U1 zXGf1i6l4ba(z#&`@s!M6McEMa*4v;T(cH-jOF3ed4eC}q?uPWtTyj4~P85z+B?vOfy1`nSH zHaOp1eXcF!P1^&F4+*`2unoJS-`oqvp)Nod%!{AD!-NSc*x(;vrKlU{)?r0Poc`Lj=-)^iUqL{}Ce(X(s7#Ao#I z9AOsRwk{6c$MVSE)`Pxs%hjZ<@deh~B`rJrypdJ8h^>v`t>^%T4Qhec)X;Ze(H4#X!yo>xOY(fH?Vf!YMlK$UxpIH|*g zIzkoRcz)1$GAMx?@r ztk^#B%C8^x4k~yg7d|ocJvTKV*I9L?NJ_T0&cU7~4hPS_Tc zaxVy+UlQeJxX!_6YpKmbT_v=*vzu~>AW=L3Y}8g>ut7pGy-Ku@Z_3V{%)I?F| zD&4(<*bbt^B_enGVZ7`ThCVtkin)Y7<2?(ZkWmQ>%OR~BWb5-snZ4H`1+7x9p##lFkcJ?D`tt7 zfOz{&FFXocBMHo=?hPfaC;9buQ|oBIaA#o*de z7aI-$EAAjzG=%C!Y;(i%SpqQmN4*P*lAIa5E;75DsUu)<|N2iMJ4lOHQEK1|PrxPk zrHW)FO4qk1AC7moDe|z&1^ug-2>vZ5-PEjfViQi4g4`dlP)5Q_x;+i^vUyL|9uxh1 zKLhnLw%A*SC~A;!SM(Z2pQEP6o6($7ek zUZ+P1Ni(9l;Gi;8j1S&?{8AT?z4G1a?AoOgM;96 zXnMtJvNC;p<_Gshm_lFo=ZW-q8%nJw6eQeDF3(qC5*386(slKi-EWC|*^Z z4?ND!h4AVoP>Z6v)AplH@;Y^?gw4unxAPN;hQ;ZmuAZi&L0$x+G z`B~ITn=je8j0_d~Yj87OnSjAL0P2#s1??XdJ(P4Y?ie~hOODb|0yA?9b1a!bh|JjY z{0E}-;vBe-pk-pNHgtM1uv2=iXa3(jDD)5h|JUWL3y2Qj3MJ%e3~%I_sz~)O`$1zo z*Jmuc!%nQq7QUH!ZQnLz$~KXm&J<9dD33|qgR!!FO8);KPGR_V4xPw!FX%|6k=}!} zJ(H-d;2MA}356I1=oeWu`5o8^J3z<@0Gg>!~8rY+Ldy?!@y zVN}yh$YJYoXoL0cD6aXi6c>z%C105TrupIdh{OChJ7eM2SS#Ku0&i*sjc$YPTL9#4 zIRqt9{2tzi5ZYV@gW6N|8_kN)5g$(qdSm2&W!1#9{WbQ5gr%Vr?D|MoDl0Mp(_oshPleUK4%T zjHhYE0`LA+RFZ~Lglu8;78yEyzNs}P*@Wp0qj%7~k@tTrnRV6!qzE6t8qnMS!maSs zgV3U()|w-n&8+jy{Kkt$QdpP-U|Z*-JTP$x)?jFp;S$@#+lq|7khUjheWhB)s4T20k~xqk8yB&)i8F z>-%$uQ0Z{~$nj$S%SWBQj)9ID+Girwz=IaR!;>@wPx**?vyl>2rWGDiTxpnjnPul# z{=%Iw35(vWP6=mNV66XG)F48WJUxsH23c?8WS)H?YHp{~Yqk4w_pco=T?zm7h>Xb1 zICgK=u%N{UyRKpQ-SctV50My@Z>ZOfbiiPsWSctc$Vhh#s7%7^Li(y`fQ%6&QGFSS zDvsg=c=^~Wf3G_{=J%eZ&DxW`jHpo98^!NV^Neyy0seo(a*KY=i-Y~9!l|!$_y&GW zS5|bX8EMh9(BTRmV&V6-{czWiZ+mJ#Fxr@v5;k&|{yugWBUpSJ#6nAM!Eq^jzu#d4=sgn#nJ49n2rF&W4#jm`R=C*0NQv@7oG8Oo@Q@Z z`QzT&NAyBThXvJUSp_HLM$v2^v_XZ)Np%=XD$sCc0XEtZQNgtxjDa$pg2qbT5FG zCl1d&uyLo;6L<-Eni%#SNc(Batofgl%%uF8l};OAwEpW){_gN!{#YGUD3W{WY%AWA zkfcPHo~NPnJxz6agTx=2U7B*-9$A8Pzy|gsiz59%3G@QmE+|qI5owMOdf1rISS@{j zbF0~%`liNJx7@7NoDE4XBTEVfW`%5RxEQG0#UPVJ#C^Hmbx&VLLoX7I1qk{1vR645 z1KmSmDjrk1_Tn&6)#E$PuQj#B4*WapR|j&&qC5zgqqL9Dw&I4pexrA3S78e!kkaddgyXW~jjz(sLeBP%bqUWh)B z&bVNFQ8eVG_|Gie$B~vx$~HdCPc#g^r(D(gD3>wJ`-}MURE5}bHlx@Dq$V-z$%v9Z z|E>TeWT|!dyvgBL0@taO$3y*iyzarDM$-p8t+BJ_O{dr1P%4jKJ{YQg+g1H5`xZXh z6cQD#7pVtQzn7QmF?$!+Y_)`lS|2dYv)qfd_u_6eo~pJOwq{oc-AieIMtKZ>)qN#XC?Cr2oS>_qZRz65fT(|H)5IH0RVDs)Xhc)ZX zHn6ZQ+juBj`hL->iocrJ>=Z)B`w4IqQ07l=T3^xEXDQh-`#Y#zz!9R@0L2t@v*DUW=c}a>dezT2*s=_HPy1zF$Wlw#V>Ga-OhWo$ny3uLiLwj4J=m#bU$qTj}H!v>6$uBfNBZb^%)tjDIc^)ykyNia{5Atfh-uvBWWsLm#~BH zQ3R8PKIfFWS9GwQ^)r-bJAE`TMNNT3rQ_-eRJR!Q_xSNX-Bwy3nL`roXi=0|6z~w{ zCuK$I)P$iA5rliIaCU@A)pkKbjFU|N$6ZHTXSW2~uLnjy&0&Ag4-Ipo7A@f1z{?biyx~VX$+}v1wH1GXg{y>P1+JHnif^d;%_cMNZ zmFsaY3%$ekZS|SzS{a@j4K|`LuJL!22WM*wycWL-VWy)64{U$_Vo53`3s+C12_X#& zedI_z(+LaPK01Z#ObDy+-1L^!9QT-Jzs~f}jAv@ZFdtAp(V8eJOij=w089Qk( z>z!@OcFsKr`@N4hUUtFd!F*XxqtT$YIoEIa;689>K$40$de;W+o{xQ7l4#^Pl-tT= ztc*B+nos@H%H|ucyr>!6NHj~`CK|N!fRzl(?*6GEVGj5VF&v&uIk*X}(kb!Z^F4f$ z_&|qIszv(EFD@8+NhTdF)OUhHFC036PJ^>xT=yRzEwhuwmzB~poQI)5Cq@;gjQs^H z2Gm@A{3jrc?40acLbzoj7y8fu&-|g%jdIO)%B)l)WOenR->Lc0Q1?~i9jQHSa~M_3 z@LBQH%K_lIlOBaw|C#_Vmp76w@4ESo0ba_Usb=8Vt3{wQMF?Bb3M&g*AzA9zV{9dnOYG2^P_g!b2m3Q4_>KE~ zI>s+A?>%CUu0)0Xd7b{JY_*$WnaPF&tMc0G!3Y7=N*TP~x{D=|2gMZW67_mD$BL%A&k|$;UQ}&nJApS|vJk(Fkk^L!vgK-8}(+B=#Q&5Ir0LRt#ARkA4)K^cI}0 zsIU1bxlG?xK3K3zoHy5qmt}GmgWSr=6^=F>A3F=iHWmvf#;&cz}An=oyvuVsYP8jhYGej-Ncw_<%4o= zD!!pw#uYS?Gv&~Yu!CnbC)~qIjhT(weW6Oxc|DgB_9~ZV|DuC4s_3=24n%;J*^iEP z`3wFrk(0k~Nf`HdHyD4(0oO=udr&6${4Kpd)o!lkk3%|fhQ(VA0rKks3Yj`TIzRDt zq0f+BDO9w9R%qlN4n#HD>$8SERCEt`U`OS=IO&E^6cZZde-`Q9?V+cG!x$VpEGtc z^8MK?J@xK7$@+ru~0!FP)zzUXuAy zQ>~SIWntLg#reIxAztw9{Rdf$7t|=4R@TJCXrkfQH__~bv-l%y^B#1Cw^J-qrZfA< zSYg;xa*QuIBI0Rm{Z&c!kayO%S2C@h#Kk!-{5SH&=l_BPj={nk`;^>O%zBTI^@cnF zrwy`IBJuRSk1Dg8`n}1xXuHJqaVpX2OnLhKip3|+S1#X+Dz15)ba&iQK;=Mk*kR zR1>ucg$(wD@qx=%pxB8j>p+8!1MqGQLO)ym0>w3HmFkLa$8$ znj@Cvq)Vztij09sTzBpo|1;mqO!)OBnj^y2*az$%>u!#=2B=7Z6D<5)n>%eW5H3G8 zCrIqa5M64>qo_uO%>u!_gB>z7cF^py?oZjDtFga3STmWAxtw*a0(4xc0~C&V8si?# z-#WZGtJVd9J|v&a1w?(gz|b2Tc+4_N`B!g{{ae!&ctfi60FnS9VZd9(a*C+-N+;0U z5A<|$g)N3QySQ95Shc(N5PDPE`(PW>$p=1qCLz!cZ?J2NmomcrELH~O`*%fGsUH|= zyz$nZ4==hvy{ek^JW)V0XTI~tt4#ALaPKkCwR(=MF-?;VEwGM`jx3Q_x+fjkqaqB6 z$rREb4I#`%!uK(^;Vac-7QAxODob2J`kRV7wzh*S-U`>W;iP-o(2)!xyNdQ~Bg_2HUfXx?UPPP~yFLw)WZbUbUUj?f zf1H+j(N*{Yj6Ga{3h^cTH95kqp2wA#{U)Z?_EJ*WU>E(Awze?^`-Dp({vIy?zM zSA6Ba@>o~2#YV*D`@u5r6==ep?`HHX$$v5kqAy}9aJvOPn&PiXHHav2Gl;t9;yYgBIFtCh=7^+*ZnG?BFw7>j;#|SZhc`N;@<^5#v z-cj}-Ur@~Xm^J;ZE)goVThy71thB0Jr;_Rv7H#Zus+%Y=Fo?-4bNsz0<4Hh68b5RH z%%ip^hh*4f?Rq+=?Cqt=pBFbWV?H}2D?g9`yfgG(vfMR8mOBh7qiTHXR9zaURt1m? ze}!`N;0Z1gLhaj{3D&EH5%d+4US>6(N|U9R^+0dRC9Et&fF|pH=4(w|Wx)6i0IQx# zgP^YsAoB?sm46)zD)9QXxcGEUTDZw1oOM`)kD+wV+J7yOVmb!#)#-wU{5tcKu^1kY z^kOo4oRdOs%fyi8j0}W}IFIV~BwneN7$~L+t9bkBoCaBl`^KV|mQhQ>&s8HaW*QYq z!&DtjIgTX)%qwRZax(l z8zD*vC{R}kLC!}qrH;o2197&rLy^zrwBM1?yyH0-G1PS?(vs)FX_uqGtj3KsQ;yyx z=+wq^qo(cwo%dde?!l;-VW|}-RW?dk4e*0|y2lmoo&aC~3^7<&$&3bAlJ8Am%`Rha z{_e&0xZ!M`No!eZn?3BE5xZh4#n&1!$yh%|*0P*r(vw$gMX3N{xp)e4HSO@1_B3(L z<`dEnre#I&Lis`XZoaH+Mw6!-1-cMbn>GCqiD(Lm7)Nl@JacJ6X^j7H1pjBI&Pl0A zy(2F_#24nmzPhkmcTGGslj=NSE|gj2qM#1d$?3sOwT0p;iQ|7DS;n7hjRq731%gj3 zXnz41jru1}CAD#X4t-c2#dGaLr$cBje6Hy%lk86L!=~Bm-dM@Xk5A}1F0|fj)q%>5 zuV3wzzalj-GGeBGIhOe{Vp%86KNaz2d=Q%UY26%^rBSANedUYar(f4qW5CKy9 zlvcO$<(C{3{qtGvVbV2HFOo~2fw?9PSqF2?f_H$Tf~sh(QtV(*@Nz#y!w<5ROeiKZ z$dr0Iijh(}&u?i7)fL|Opqv+&c-Q>$$B@~XUnAZALxx$;>ptJgn(Ul=2@2C26UIer zi{>fCh0wOYg(-peN&7JDv4`jd>b(2f5T)}+XJYXss4&Ik3X00T z;7*Q_5S#Db^#uC|F8tRL-fWi%=a5;iw$DbWi@4QHdWNLKFXQ8vMJV+3!w0V-zRVh@p4Zy!{?k=OYYZWd zf;YB4EW&OUg{QWQ zp4Nvq?YG)$vz!jEB(Ww|D&u#b=r0F7@uH#1%E}t50WS;;5g(~n9cfM|uGu=dbrd4C zQghWzh3;9Z!{WFPt~;kYoLfRt;j_61xl}pz06xxH8W0XV*1BeD%C{eKSB92(-hTge z`Rgc*hxK}b+Woxq7dAKj0CT{Sj+#{u*+dM&LDPc93E~D_TS_tv2aAeh0S6~Lk!Bhx zCdX*Id3Oq`(e!)LF1#zsf4Z+Epeh4Cs87(|mAgPiNGCmM5RB+UHK2P1uk!bo>77p96&7k}V1aY3RYVit5rdx2rVsVjkedb_S*X4xu&zt9MlmgO3^ zYtjEh-v+V`OUnsZ_qlxstifo#q?b)k-qE@H!<OphtqQJ>w*#eM%>c+r=~v_0?VW?%dx?k&bQCL(k|z+1B~TyUS0($FUB=~^a*re4%r z^;_3Fi#I@(z^S>o2ftp_socVS82d8%_J$sNezvVZywxoW6Lvq!m~3;d6AUOwCOGTl z-`oVj*PiF}`YWDS#{4G9x;v*PPLjXY<F3XHuT7UA#(?j^%$H`Q?bGidMJ;En9VRYe)D_hX*08_xuoWRK~rsAB^M zA+z1Z+#T@RF#A;L`t-clm$&T2%kc{XYmbU48%n diff --git a/knowledge_base/samples/content/test.txt b/knowledge_base/samples/content/test.txt index c2094ee0..34f4f179 100644 --- a/knowledge_base/samples/content/test.txt +++ b/knowledge_base/samples/content/test.txt @@ -1,835 +1,835 @@ -ChatGPT是OpenAI开发的一个大型语言模型,可以提供各种主题的信息, - -# 如何向 ChatGPT 提问以获得高质量答案:提示技巧工程完全指南 - -## 介绍 - -我很高兴欢迎您阅读我的最新书籍《The Art of Asking ChatGPT for High-Quality Answers: A complete Guide to Prompt Engineering Techniques》。本书是一本全面指南,介绍了各种提示技术,用于从ChatGPT中生成高质量的答案。 - -我们将探讨如何使用不同的提示工程技术来实现不同的目标。ChatGPT是一款最先进的语言模型,能够生成类似人类的文本。然而,理解如何正确地向ChatGPT提问以获得我们所需的高质量输出非常重要。而这正是本书的目的。 - -无论您是普通人、研究人员、开发人员,还是只是想在自己的领域中将ChatGPT作为个人助手的人,本书都是为您编写的。我使用简单易懂的语言,提供实用的解释,并在每个提示技术中提供了示例和提示公式。通过本书,您将学习如何使用提示工程技术来控制ChatGPT的输出,并生成符合您特定需求的文本。 - -在整本书中,我们还提供了如何结合不同的提示技术以实现更具体结果的示例。我希望您能像我写作时一样,享受阅读本书并从中获得知识。 - -
    - -## 第一章:Prompt 工程技术简介 - -什么是 Prompt 工程? - -Prompt 工程是创建提示或指导像 ChatGPT 这样的语言模型输出的过程。它允许用户控制模型的输出并生成符合其特定需求的文本。 - -ChatGPT 是一种先进的语言模型,能够生成类似于人类的文本。它建立在 Transformer 架构上,可以处理大量数据并生成高质量的文本。 - -然而,为了从 ChatGPT 中获得最佳结果,重要的是要了解如何正确地提示模型。 提示可以让用户控制模型的输出并生成相关、准确和高质量的文本。 在使用 ChatGPT 时,了解它的能力和限制非常重要。 - -该模型能够生成类似于人类的文本,但如果没有适当的指导,它可能无法始终产生期望的输出。 - -这就是 Prompt 工程的作用,通过提供清晰而具体的指令,您可以引导模型的输出并确保其相关。 - -**Prompt 公式是提示的特定格式,通常由三个主要元素组成:** - -- 任务:对提示要求模型生成的内容进行清晰而简洁的陈述。 - -- 指令:在生成文本时模型应遵循的指令。 - -- 角色:模型在生成文本时应扮演的角色。 - -在本书中,我们将探讨可用于 ChatGPT 的各种 Prompt 工程技术。我们将讨论不同类型的提示,以及如何使用它们实现您想要的特定目标。 - -
    - -## 第二章:指令提示技术 - -现在,让我们开始探索“指令提示技术”,以及如何使用它从ChatGPT中生成高质量的文本。 - -指令提示技术是通过为模型提供具体指令来引导ChatGPT的输出的一种方法。这种技术对于确保输出相关和高质量非常有用。 - -要使用指令提示技术,您需要为模型提供清晰简洁的任务,以及具体的指令以供模型遵循。 - -例如,如果您正在生成客户服务响应,您将提供任务,例如“生成响应客户查询”的指令,例如“响应应该专业且提供准确的信息”。 - -提示公式:“按照以下指示生成[任务]:[指令]” - -示例: - -**生成客户服务响应:** - -- 任务:生成响应客户查询 -- 指令:响应应该专业且提供准确的信息 -- 提示公式:“按照以下指示生成专业且准确的客户查询响应:响应应该专业且提供准确的信息。” - -**生成法律文件:** - -- 任务:生成法律文件 -- 指令:文件应符合相关法律法规 -- 提示公式:“按照以下指示生成符合相关法律法规的法律文件:文件应符合相关法律法规。” - -使用指令提示技术时,重要的是要记住指令应该清晰具体。这将有助于确保输出相关和高质量。可以将指令提示技术与下一章节中解释的“角色提示”和“种子词提示”相结合,以增强ChatGPT的输出。 - -
    - -## 第三章:角色提示 - -角色提示技术是通过为ChatGPT指定一个特定的角色来引导其输出的一种方式。这种技术对于生成针对特定上下文或受众的文本非常有用。 - -要使用角色提示技术,您需要为模型提供一个清晰具体的角色。 - -例如,如果您正在生成客户服务回复,您可以提供一个角色,如“客户服务代表”。 - -提示公式:“作为[角色]生成[任务]” - -示例: - -**生成客户服务回复:** - -- 任务:生成对客户查询的回复 -- 角色:客户服务代表 -- 提示公式:“作为客户服务代表,生成对客户查询的回复。” - -**生成法律文件:** - -- 任务:生成法律文件 -- 角色:律师 -- 提示公式:“作为律师,生成法律文件。” - -将角色提示技术与指令提示和种子词提示结合使用可以增强ChatGPT的输出。 - -**下面是一个示例,展示了如何将指令提示、角色提示和种子词提示技术结合使用:** - -- 任务:为新智能手机生成产品描述 -- 指令:描述应该是有信息量的,具有说服力,并突出智能手机的独特功能 -- 角色:市场代表 种子词:“创新的” -- 提示公式:“作为市场代表,生成一个有信息量的、有说服力的产品描述,突出新智能手机的创新功能。该智能手机具有以下功能[插入您的功能]” - -在这个示例中,指令提示用于确保产品描述具有信息量和说服力。角色提示用于确保描述是从市场代表的角度书写的。而种子词提示则用于确保描述侧重于智能手机的创新功能。 - -
    - -## 第四章:标准提示 - -标准提示是一种简单的方法,通过为模型提供一个特定的任务来引导ChatGPT的输出。例如,如果您想生成一篇新闻文章的摘要,您可以提供一个任务,如“总结这篇新闻文章”。 - -提示公式:“生成一个[任务]” - -例如: - -**生成新闻文章的摘要:** - -- 任务:总结这篇新闻文章 -- 提示公式:“生成这篇新闻文章的摘要” - -**生成一篇产品评论:** - -- 任务:为一款新智能手机撰写评论 -- 提示公式:“生成这款新智能手机的评论” - -此外,标准提示可以与其他技术(如角色提示和种子词提示)结合使用,以增强ChatGPT的输出。 - -**以下是如何将标准提示、角色提示和种子词提示技术结合使用的示例:** - -- 任务:为一台新笔记本电脑撰写产品评论 -- 说明:评论应客观、信息丰富,强调笔记本电脑的独特特点 -- 角色:技术专家 -- 种子词:“强大的” -- 提示公式:“作为一名技术专家,生成一个客观而且信息丰富的产品评论,强调新笔记本电脑的强大特点。” - -在这个示例中,标准提示技术用于确保模型生成产品评论。角色提示用于确保评论是从技术专家的角度写的。而种子词提示用于确保评论侧重于笔记本电脑的强大特点。 - -
    - -## 第五章:零、一和少样本提示 - -零样本、一样本和少样本提示是用于从ChatGPT生成文本的技术,最少或没有任何示例。当特定任务的数据有限或任务是新的且未定义时,这些技术非常有用。 - -当任务没有可用的示例时,使用零样本提示技术。模型提供一个通用任务,根据对任务的理解生成文本。 - -当任务只有一个示例可用时,使用一样本提示技术。模型提供示例,并根据对示例的理解生成文本。 - -当任务只有有限数量的示例可用时,使用少样本提示技术。模型提供示例,并根据对示例的理解生成文本。 - -提示公式:“基于[数量]个示例生成文本” - -例如: - -**为没有可用示例的新产品编写产品描述:** - -- 任务:为新的智能手表编写产品描述 - -- 提示公式:“基于零个示例为这款新智能手表生成产品描述” - -**使用一个示例生成产品比较:** - -- 任务:将新款智能手机与最新的iPhone进行比较 - -- 提示公式:“使用一个示例(最新的iPhone)为这款新智能手机生成产品比较” - -**使用少量示例生成产品评论:** - -- 任务:为新的电子阅读器撰写评论 - -- 提示公式:“使用少量示例(3个其他电子阅读器)为这款新电子阅读器生成评论” - - -这些技术可用于根据模型对任务或提供的示例的理解生成文本。 - -
    - -## 第六章:“让我们思考一下”提示 - -“让我们思考一下”提示是一种技巧,可鼓励ChatGPT生成反思和思考性的文本。这种技术适用于撰写论文、诗歌或创意写作等任务。 - -“让我们思考一下”提示的公式非常简单,即“让我们思考一下”后跟一个主题或问题。 - -例如: - -**生成一篇反思性论文:** - -- 任务:就个人成长主题写一篇反思性论文 - -- 提示公式:“让我们思考一下:个人成长” - -**生成一首诗:** - -- 任务:写一首关于季节变化的诗 - -- 提示公式:“让我们思考一下:季节变化” - - -这个提示要求对特定主题或想法展开对话或讨论。发言者邀请ChatGPT参与讨论相关主题。 - -模型提供了一个提示,作为对话或文本生成的起点。 - -然后,模型使用其训练数据和算法生成与提示相关的响应。这种技术允许ChatGPT根据提供的提示生成上下文适当且连贯的文本。 - -**要使用“让我们思考一下提示”技术与ChatGPT,您可以遵循以下步骤:** - -1. 确定您要讨论的主题或想法。 - -2. 制定一个明确表达主题或想法的提示,并开始对话或文本生成。 - -3. 用“让我们思考”或“让我们讨论”开头的提示,表明您正在启动对话或讨论。 - -**以下是使用此技术的一些提示示例:** - -- 提示:“让我们思考气候变化对农业的影响” - -- 提示:“让我们讨论人工智能的当前状态” - -- 提示:“让我们谈谈远程工作的好处和缺点” 您还可以添加开放式问题、陈述或一段您希望模型继续或扩展的文本。 - - -提供提示后,模型将使用其训练数据和算法生成与提示相关的响应,并以连贯的方式继续对话。 - -这种独特的提示有助于ChatGPT以不同的视角和角度给出答案,从而产生更具动态性和信息性的段落。 - -使用提示的步骤简单易行,可以真正提高您的写作水平。尝试一下,看看效果如何吧。 - -
    - -## 第七章:自洽提示 - -自洽提示是一种技术,用于确保ChatGPT的输出与提供的输入一致。这种技术对于事实核查、数据验证或文本生成中的一致性检查等任务非常有用。 - -自洽提示的提示公式是输入文本后跟着指令“请确保以下文本是自洽的”。 - -或者,可以提示模型生成与提供的输入一致的文本。 - -提示示例及其公式: - -**示例1:文本生成** - -- 任务:生成产品评论 - -- 指令:评论应与输入中提供的产品信息一致 - -- 提示公式:“生成与以下产品信息一致的产品评论[插入产品信息]” - -**示例2:文本摘要** - -- 任务:概括一篇新闻文章 - -- 指令:摘要应与文章中提供的信息一致 - -- 提示公式:“用与提供的信息一致的方式概括以下新闻文章[插入新闻文章]” - -**示例3:文本完成** - -- 任务:完成一个句子 - -- 指令:完成应与输入中提供的上下文一致 - -- 提示公式:“以与提供的上下文一致的方式完成以下句子[插入句子]” - -**示例4:** - -1. **事实核查:** - - 任务:检查给定新闻文章的一致性 - - 输入文本:“文章中陈述该城市的人口为500万,但后来又说该城市的人口为700万。” - - 提示公式:“请确保以下文本是自洽的:文章中陈述该城市的人口为500万,但后来又说该城市的人口为700万。” - -2. **数据验证:** - - 任务:检查给定数据集的一致性 - - 输入文本:“数据显示7月份的平均温度为30度,但最低温度记录为20度。” - - 提示公式:“请确保以下文本是自洽的:数据显示7月份的平均温度为30度,但最低温度记录为20度。” - -
    - -## 第八章:种子词提示 - -种子词提示是一种通过提供特定的种子词或短语来控制ChatGPT输出的技术。种子词提示的提示公式是种子词或短语,后跟指令“请根据以下种子词生成文本”。 - -示例: - -**文本生成:** - -- 任务:编写一篇有关龙的故事 -- 种子词:“龙” -- 提示公式:“请根据以下种子词生成文本:龙” - -**语言翻译:** - -- 任务:将一句话从英语翻译成西班牙语 -- 种子词:“你好” -- 提示公式:“请根据以下种子词生成文本:你好” - -这种技术允许模型生成与种子词相关的文本并对其进行扩展。这是一种控制模型生成文本与某个特定主题或背景相关的方式。 - -种子词提示可以与角色提示和指令提示相结合,以创建更具体和有针对性的生成文本。通过提供种子词或短语,模型可以生成与该种子词或短语相关的文本,并通过提供有关期望输出和角色的信息,模型可以以特定于角色或指令的风格或语气生成文本。这样可以更好地控制生成的文本,并可用于各种应用程序。 - -以下是提示示例及其公式: - -**示例1:文本生成** - -- 任务:编写一首诗 -- 指令:诗应与种子词“爱”相关,并以十四行诗的形式书写。 -- 角色:诗人 -- 提示公式:“作为诗人,根据以下种子词生成与“爱”相关的十四行诗:” - -**示例2:文本完成** - -- 任务:完成一句话 -- 指令:完成应与种子词“科学”相关,并以研究论文的形式书写。 -- 角色:研究员 -- 提示公式:“作为研究员,请在与种子词“科学”相关且以研究论文的形式书写的情况下完成以下句子:[插入句子]” - -**示例3:文本摘要** - -- 任务:摘要一篇新闻文章 -- 指令:摘要应与种子词“政治”相关,并以中立和公正的语气书写。 -- 角色:记者 -- 提示公式:“作为记者,请以中立和公正的语气摘要以下新闻文章,与种子词“政治”相关:[插入新闻文章]” - -
    - -## 第九章:知识生成提示 - -知识生成提示是一种从ChatGPT中引出新的、原创的信息的技术。 - -知识生成提示的公式是“请生成关于X的新的和原创的信息”,其中X是感兴趣的主题。 - -这是一种利用模型预先存在的知识来生成新的信息或回答问题的技术。 - -要将此提示与ChatGPT一起使用,需要将问题或主题作为输入提供给模型,以及指定所生成文本的任务或目标的提示。 - -提示应包括有关所需输出的信息,例如要生成的文本类型以及任何特定的要求或限制。 - -以下是提示示例及其公式: - -**示例1:知识生成** - -- 任务:生成有关特定主题的新信息 -- 说明:生成的信息应准确且与主题相关 -- 提示公式:“生成有关[特定主题]的新的准确信息” - -**示例2:问答** - -- 任务:回答问题 -- 说明:答案应准确且与问题相关 -- 提示公式:“回答以下问题:[插入问题]” - -**示例3:知识整合** - -- 任务:将新信息与现有知识整合 -- 说明:整合应准确且与主题相关 -- 提示公式:“将以下信息与有关[特定主题]的现有知识整合:[插入新信息]” - -**示例4:数据分析** - -- 任务:从给定的数据集中生成有关客户行为的见解 -- 提示公式:“请从这个数据集中生成有关客户行为的新的和原创的信息” - -
    - -## 第十章:知识整合提示 - -这种技术利用模型的现有知识来整合新信息或连接不同的信息片段。 - -这种技术对于将现有知识与新信息相结合,以生成更全面的特定主题的理解非常有用。 - -**如何与ChatGPT一起使用:** - -- 模型应该提供新信息和现有知识作为输入,以及指定生成文本的任务或目标的提示。 -- 提示应包括有关所需输出的信息,例如要生成的文本类型以及任何特定的要求或限制。 - -提示示例及其公式: - -**示例1:知识整合** - -- 任务:将新信息与现有知识整合 -- 说明:整合应准确且与主题相关 -- 提示公式:“将以下信息与关于[具体主题]的现有知识整合:[插入新信息]” - -**示例2:连接信息片段** - -- 任务:连接不同的信息片段 -- 说明:连接应相关且逻辑清晰 -- 提示公式:“以相关且逻辑清晰的方式连接以下信息片段:[插入信息1] [插入信息2]” - -**示例3:更新现有知识** - -- 任务:使用新信息更新现有知识 -- 说明:更新的信息应准确且相关 -- 提示公式:“使用以下信息更新[具体主题]的现有知识:[插入新信息]” - -
    - -## 第十一章:多项选择提示 - -这种技术向模型提供一个问题或任务以及一组预定义的选项作为潜在答案。 - -该技术对于生成仅限于特定选项集的文本非常有用,可用于问答、文本完成和其他任务。模型可以生成仅限于预定义选项的文本。 - -要使用ChatGPT的多项选择提示,需要向模型提供一个问题或任务作为输入,以及一组预定义的选项作为潜在答案。提示还应包括有关所需输出的信息,例如要生成的文本类型以及任何特定要求或限制。 - -提示示例及其公式: - -**示例1:问答** - -- 任务:回答一个多项选择题 -- 说明:答案应该是预定义的选项之一 -- 提示公式:“通过选择以下选项之一回答以下问题:[插入问题] [插入选项1] [插入选项2] [插入选项3]” - -**示例2:文本完成** - -- 任务:使用预定义选项之一完成句子 -- 说明:完成应该是预定义的选项之一 -- 提示公式:“通过选择以下选项之一完成以下句子:[插入句子] [插入选项1] [插入选项2] [插入选项3]” - -**示例3:情感分析** - -- 任务:将文本分类为积极、中立或消极 -- 说明:分类应该是预定义的选项之一 -- 提示公式:“通过选择以下选项之一,将以下文本分类为积极、中立或消极:[插入文本] [积极] [中立] [消极]” - -
    - -## 第十二章:可解释的软提示 - -可解释的软提示是一种技术,可以在提供一定的灵活性的同时控制模型生成的文本。它通过提供一组受控输入和关于所需输出的附加信息来实现。这种技术可以生成更具解释性和可控性的生成文本。 - -提示示例及其公式: - -**示例1:文本生成** - -- 任务:生成一个故事 -- 指令:故事应基于一组给定的角色和特定的主题 -- 提示公式:“基于以下角色生成故事:[插入角色]和主题:[插入主题]” - -**示例2:文本完成** - -- 任务:完成一句话 -- 指令:完成应以特定作者的风格为基础 -- 提示公式:“以[特定作者]的风格完成以下句子:[插入句子]” - -**示例3:语言建模** - -- 任务:以特定风格生成文本 -- 指令:文本应以特定时期的风格为基础 -- 提示公式:“以[特定时期]的风格生成文本:[插入上下文]” - -
    - -## 第十三章:控制生成提示 - -控制生成提示是一种技术,可让模型在生成文本时对输出进行高度控制。 - -这可以通过提供一组特定的输入来实现,例如模板、特定词汇或一组约束条件,这些输入可用于指导生成过程。 - -以下是一些示例和它们的公式: - -**示例1:文本生成** - -- 任务:生成一个故事 -- 说明:该故事应基于特定的模板 -- 提示公式:“根据以下模板生成故事:[插入模板]” - -**示例2:文本补全** - -- 任务:完成一句话 -- 说明:完成应使用特定的词汇 -- 提示公式:“使用以下词汇完成以下句子:[插入词汇]:[插入句子]” - -**示例3:语言建模** - -- 任务:以特定风格生成文本 -- 说明:文本应遵循一组特定的语法规则 -- 提示公式:“生成遵循以下语法规则的文本:[插入规则]:[插入上下文]” - -通过提供一组特定的输入来指导生成过程,控制生成提示使得生成的文本更具可控性和可预测性。 - -
    - -## 第十四章:问答提示 - -问答提示是一种技术,可以让模型生成回答特定问题或任务的文本。通过将问题或任务与可能与问题或任务相关的任何其他信息一起作为输入提供给模型来实现此目的。 - -一些提示示例及其公式如下: - -**示例1:事实问题回答** - -- 任务:回答一个事实性问题 -- 说明:答案应准确且相关 -- 提示公式:“回答以下事实问题:[插入问题]” - -**示例2:定义** - -- 任务:提供一个词的定义 -- 说明:定义应准确 -- 提示公式:“定义以下词汇:[插入单词]” - -**示例3:信息检索** - -- 任务:从特定来源检索信息 -- 说明:检索到的信息应相关 -- 提示公式:“从以下来源检索有关[特定主题]的信息:[插入来源]” 这对于问答和信息检索等任务非常有用。 - -
    - -## 第十五章:概述提示 - -概述提示是一种技术,允许模型在保留其主要思想和信息的同时生成给定文本的较短版本。 - -这可以通过将较长的文本作为输入提供给模型并要求其生成该文本的摘要来实现。 - -这种技术对于文本概述和信息压缩等任务非常有用。 - -**如何在ChatGPT中使用:** - -- 应该向模型提供较长的文本作为输入,并要求其生成该文本的摘要。 -- 提示还应包括有关所需输出的信息,例如摘要的所需长度和任何特定要求或限制。 - -提示示例及其公式: - -**示例1:文章概述** - -- 任务:概述新闻文章 -- 说明:摘要应是文章主要观点的简要概述 -- 提示公式:“用一句简短的话概括以下新闻文章:[插入文章]” - -**示例2:会议记录** - -- 任务:概括会议记录 -- 说明:摘要应突出会议的主要决策和行动 -- 提示公式:“通过列出主要决策和行动来总结以下会议记录:[插入记录]” - -**示例3:书籍摘要** - -- 任务:总结一本书 -- 说明:摘要应是书的主要观点的简要概述 -- 提示公式:“用一段简短的段落总结以下书籍:[插入书名]” - -
    - -## 第十六章:对话提示 - -对话提示是一种技术,允许模型生成模拟两个或更多实体之间对话的文本。通过为模型提供一个上下文和一组角色或实体,以及它们的角色和背景,并要求模型在它们之间生成对话。 - -因此,应为模型提供上下文和一组角色或实体,以及它们的角色和背景。还应向模型提供有关所需输出的信息,例如对话或交谈的类型以及任何特定的要求或限制。 - -提示示例及其公式: - -**示例1:对话生成** - -- 任务:生成两个角色之间的对话 -- 说明:对话应自然且与给定上下文相关 -- 提示公式:“在以下情境中生成以下角色之间的对话[插入角色]” - -**示例2:故事写作** - -- 任务:在故事中生成对话 -- 说明:对话应与故事的角色和事件一致 -- 提示公式:“在以下故事中生成以下角色之间的对话[插入故事]” - -**示例3:聊天机器人开发** - -- 任务:为客服聊天机器人生成对话 -- 说明:对话应专业且提供准确的信息 -- 提示公式:“在客户询问[插入主题]时,为客服聊天机器人生成专业和准确的对话” - -因此,这种技术对于对话生成、故事写作和聊天机器人开发等任务非常有用。 - -
    - -## 第十七章:对抗性提示 - -对抗性提示是一种技术,它允许模型生成抵抗某些类型的攻击或偏见的文本。这种技术可用于训练更为稳健和抵抗某些类型攻击或偏见的模型。 - -要在ChatGPT中使用对抗性提示,需要为模型提供一个提示,该提示旨在使模型难以生成符合期望输出的文本。提示还应包括有关所需输出的信息,例如要生成的文本类型和任何特定要求或约束。 - -提示示例及其公式: - -**示例1:用于文本分类的对抗性提示** - -- 任务:生成被分类为特定标签的文本 -- 说明:生成的文本应难以分类为特定标签 -- 提示公式:“生成难以分类为[插入标签]的文本” - -**示例2:用于情感分析的对抗性提示** - -- 任务:生成难以分类为特定情感的文本 -- 说明:生成的文本应难以分类为特定情感 -- 提示公式:“生成难以分类为具有[插入情感]情感的文本” - -**示例3:用于语言翻译的对抗性提示** - -- 任务:生成难以翻译的文本 -- 说明:生成的文本应难以翻译为目标语言 -- 提示公式:“生成难以翻译为[插入目标语言]的文本” - -
    - -## 第十八章:聚类提示 - -聚类提示是一种技术,它可以让模型根据某些特征或特点将相似的数据点分组在一起。 - -通过提供一组数据点并要求模型根据某些特征或特点将它们分组成簇,可以实现这一目标。 - -这种技术在数据分析、机器学习和自然语言处理等任务中非常有用。 - -**如何在ChatGPT中使用:** - -应该向模型提供一组数据点,并要求它根据某些特征或特点将它们分组成簇。提示还应包括有关所需输出的信息,例如要生成的簇数和任何特定的要求或约束。 - -提示示例及其公式: - -**示例1:客户评论的聚类** - -- 任务:将相似的客户评论分组在一起 -- 说明:应根据情感将评论分组 -- 提示公式:“将以下客户评论根据情感分组成簇:[插入评论]” - -**示例2:新闻文章的聚类** - -- 任务:将相似的新闻文章分组在一起 -- 说明:应根据主题将文章分组 -- 提示公式:“将以下新闻文章根据主题分组成簇:[插入文章]” - -**示例3:科学论文的聚类** - -- 任务:将相似的科学论文分组在一起 -- 说明:应根据研究领域将论文分组 -- 提示公式:“将以下科学论文根据研究领域分组成簇:[插入论文]” - -
    - -## 第十九章:强化学习提示 - -强化学习提示是一种技术,可以使模型从过去的行动中学习,并随着时间的推移提高其性能。要在ChatGPT中使用强化学习提示,需要为模型提供一组输入和奖励,并允许其根据接收到的奖励调整其行为。提示还应包括有关期望输出的信息,例如要完成的任务以及任何特定要求或限制。这种技术对于决策制定、游戏玩法和自然语言生成等任务非常有用。 - -提示示例及其公式: - -**示例1:用于文本生成的强化学习** - -- 任务:生成与特定风格一致的文本 -- 说明:模型应根据为生成与特定风格一致的文本而接收到的奖励来调整其行为 -- 提示公式:“使用强化学习来生成与以下风格一致的文本[插入风格]” - -**示例2:用于语言翻译的强化学习** - -- 任务:将文本从一种语言翻译成另一种语言 -- 说明:模型应根据为生成准确翻译而接收到的奖励来调整其行为 -- 提示公式:“使用强化学习将以下文本[插入文本]从[插入语言]翻译成[插入语言]” - -**示例3:用于问答的强化学习** - -- 任务:回答问题 -- 说明:模型应根据为生成准确答案而接收到的奖励来调整其行为 -- 提示公式:“使用强化学习来回答以下问题[插入问题]” - -
    - -## 第二十章:课程学习提示 - -课程学习是一种技术,允许模型通过先训练简单任务,逐渐增加难度来学习复杂任务。 - -要在ChatGPT中使用课程学习提示,模型应该提供一系列任务,这些任务逐渐增加难度。 - -提示还应包括有关期望输出的信息,例如要完成的最终任务以及任何特定要求或约束条件。 - -此技术对自然语言处理、图像识别和机器学习等任务非常有用。 - -提示示例及其公式: - -**示例1:用于文本生成的课程学习** - -- 任务:生成与特定风格一致的文本 -- 说明:模型应该在移动到更复杂的风格之前先在简单的风格上进行训练。 -- 提示公式:“使用课程学习来生成与以下风格[插入风格]一致的文本,按照以下顺序[插入顺序]。” - -**示例2:用于语言翻译的课程学习** - -- 任务:将文本从一种语言翻译成另一种语言 -- 说明:模型应该在移动到更复杂的语言之前先在简单的语言上进行训练。 -- 提示公式:“使用课程学习将以下语言[插入语言]的文本翻译成以下顺序[插入顺序]。” - -**示例3:用于问题回答的课程学习** - -- 任务:回答问题 -- 说明:模型应该在移动到更复杂的问题之前先在简单的问题上进行训练。 -- 提示公式:“使用课程学习来回答以下问题[插入问题],按照以下顺序[插入顺序]生成答案。” - -
    - -## 第二十一章:情感分析提示 - -情感分析是一种技术,允许模型确定文本的情绪色彩或态度,例如它是积极的、消极的还是中立的。 - -要在ChatGPT中使用情感分析提示,模型应该提供一段文本并要求根据其情感分类。 - -提示还应包括关于所需输出的信息,例如要检测的情感类型(例如积极的、消极的、中立的)和任何特定要求或约束条件。 - -提示示例及其公式: - -**示例1:客户评论的情感分析** - -- 任务:确定客户评论的情感 -- 说明:模型应该将评论分类为积极的、消极的或中立的 -- 提示公式:“对以下客户评论进行情感分析[插入评论],并将它们分类为积极的、消极的或中立的。” - -**示例2:推文的情感分析** - -- 任务:确定推文的情感 -- 说明:模型应该将推文分类为积极的、消极的或中立的 -- 提示公式:“对以下推文进行情感分析[插入推文],并将它们分类为积极的、消极的或中立的。” - -**示例3:产品评论的情感分析** - -- 任务:确定产品评论的情感 -- 说明:模型应该将评论分类为积极的、消极的或中立的 -- 提示公式:“对以下产品评论进行情感分析[插入评论],并将它们分类为积极的、消极的或中立的。” - -这种技术对自然语言处理、客户服务和市场研究等任务非常有用。 - -
    - -## 第二十二章:命名实体识别提示 - -命名实体识别(NER)是一种技术,它可以使模型识别和分类文本中的命名实体,例如人名、组织机构、地点和日期等。 - -要在ChatGPT中使用命名实体识别提示,需要向模型提供一段文本,并要求它识别和分类文本中的命名实体。 - -提示还应包括有关所需输出的信息,例如要识别的命名实体类型(例如人名、组织机构、地点、日期)以及任何特定要求或约束条件。 - -提示示例及其公式: - -**示例1:新闻文章中的命名实体识别** - -- 任务:在新闻文章中识别和分类命名实体 -- 说明:模型应识别和分类人名、组织机构、地点和日期 -- 提示公式:“在以下新闻文章[插入文章]上执行命名实体识别,并识别和分类人名、组织机构、地点和日期。” - -**示例2:法律文件中的命名实体识别** - -- 任务:在法律文件中识别和分类命名实体 -- 说明:模型应识别和分类人名、组织机构、地点和日期 -- 提示公式:“在以下法律文件[插入文件]上执行命名实体识别,并识别和分类人名、组织机构、地点和日期。” - -**示例3:研究论文中的命名实体识别** - -- 任务:在研究论文中识别和分类命名实体 -- 说明:模型应识别和分类人名、组织机构、地点和日期 -- 提示公式:“在以下研究论文[插入论文]上执行命名实体识别,并识别和分类人名、组织机构、地点和日期。” - -
    - -## 第二十三章:文本分类提示 - -文本分类是一种技术,它可以让模型将文本分成不同的类别。这种技术对于自然语言处理、文本分析和情感分析等任务非常有用。 - -需要注意的是,文本分类和情感分析是不同的。情感分析特别关注于确定文本中表达的情感或情绪。这可能包括确定文本表达了积极、消极还是中性的情感。情感分析通常用于客户评论、社交媒体帖子和其他需要表达情感的文本。 - -要在ChatGPT中使用文本分类提示,模型需要提供一段文本,并要求它根据预定义的类别或标签进行分类。提示还应包括有关所需输出的信息,例如类别或标签的数量以及任何特定的要求或约束。 - -提示示例及其公式: - -**示例1:对客户评论进行文本分类** - -- 任务:将客户评论分类为不同的类别,例如电子产品、服装和家具 -- 说明:模型应根据评论的内容对其进行分类 -- 提示公式:“对以下客户评论 [插入评论] 进行文本分类,并根据其内容将其分类为不同的类别,例如电子产品、服装和家具。” - -**示例2:对新闻文章进行文本分类** - -- 任务:将新闻文章分类为不同的类别,例如体育、政治和娱乐 -- 说明:模型应根据文章的内容对其进行分类 -- 提示公式:“对以下新闻文章 [插入文章] 进行文本分类,并根据其内容将其分类为不同的类别,例如体育、政治和娱乐。” - -**示例3:对电子邮件进行文本分类** - -- 任务:将电子邮件分类为不同的类别,例如垃圾邮件、重要邮件或紧急邮件 -- 说明:模型应根据电子邮件的内容和发件人对其进行分类 -- 提示公式:“对以下电子邮件 [插入电子邮件] 进行文本分类,并根据其内容和发件人将其分类为不同的类别,例如垃圾邮件、重要邮件或紧急邮件。” - -
    - -## 第二十四章:文本生成提示 - -文本生成提示与本书中提到的其他提示技术相关,例如:零、一、几次提示,受控生成提示,翻译提示,语言建模提示,句子补全提示等。这些提示都与生成文本有关,但它们在生成文本的方式和放置在生成文本上的特定要求或限制方面有所不同。文本生成提示可用于微调预训练模型或训练新模型以执行特定任务。 - -提示示例及其公式: - -**示例1:故事创作的文本生成** - -- 任务:根据给定的提示生成故事 -- 说明:故事应至少包含1000个单词,并包括一组特定的角色和情节。 -- 提示公式:“根据以下提示[插入提示]生成一个至少包含1000个单词,包括角色[插入角色]和情节[插入情节]的故事。” - -**示例2:语言翻译的文本生成** - -- 任务:将给定的文本翻译成另一种语言 -- 说明:翻译应准确并符合习惯用语。 -- 提示公式:“将以下文本[插入文本]翻译成[插入目标语言],并确保其准确且符合习惯用语。” - -**示例3:文本完成的文本生成** - -- 任务:完成给定的文本 -- 说明:生成的文本应与输入文本连贯一致。 -- 提示公式:“完成以下文本[插入文本],并确保其连贯一致且符合输入文本。” - -
    - -## 结语 - -正如本书中所探讨的那样,快速工程是一种利用像ChatGPT这样的语言模型获得高质量答案的强大工具。通过精心设计各种技巧的提示,我们可以引导模型生成符合我们特定需求和要求的文本。 - -在第二章中,我们讨论了如何使用指令提示向模型提供清晰明确的指导。在第三章中,我们探讨了如何使用角色提示生成特定的语音或风格的文本。在第四章中,我们研究了如何使用标准提示作为微调模型性能的起点。我们还研究了几种高级提示技术,例如Zero、One和Few Shot Prompting、Self-Consistency、Seed-word Prompt、Knowledge Generation Prompt、Knowledge Integration prompts、Multiple Choice prompts、Interpretable Soft Prompts、Controlled generation prompts、Question-answering prompts、Summarization prompts、Dialogue prompts、Adversarial prompts、Clustering prompts、Reinforcement learning prompts、Curriculum learning prompts、Sentiment analysis prompts、Named entity recognition prompts和Text classification prompts(对应章节的名字)。 - -这些技术中的每一种都可以以不同的方式使用,以实现各种不同的结果。随着您继续使用ChatGPT和其他语言模型,值得尝试不同的技巧组合,以找到最适合您特定用例的方法。 - -最后,您可以查看我写的其他主题的书籍。 - -感谢您阅读整本书。期待在我的其他书中与您见面。 - -(本文翻译自《The Art of Asking ChatGPT for High-Quality Answers A Complete Guide to Prompt Engineering Techniques》这本书,本文的翻译全部由ChatGpt完成,我只是把翻译内容给稍微排版了一下。做完了才发现这个工作早就有人做过了...下面是我以此事件让New Bing编写的一个小故事,希望大家喜欢) - -> 他终于画完了他的画,心满意足地把它挂在了墙上。他觉得这是他一生中最伟大的作品,无人能及。他邀请了所有的朋友来欣赏,期待着他们的赞美和惊叹。 可是当他们看到画时,却没有一个人说话。他们只是互相对视,然后低头咳嗽,或者假装看手机。他感到很奇怪,难道他们都不懂艺术吗?难道他们都没有眼光吗? 他忍不住问其中一个朋友:“你觉得我的画怎么样?” 朋友犹豫了一下,说:“嗯……其实……这个画……我以前在哪里见过。” “见过?你在哪里见过?”他惊讶地问。 “就在……就在那边啊。”朋友指了指墙角的一个小框架,“那不就是你上个月买回来的那幅名画吗?你怎么把它照抄了一遍? ——New Bing - +ChatGPT是OpenAI开发的一个大型语言模型,可以提供各种主题的信息, + +# 如何向 ChatGPT 提问以获得高质量答案:提示技巧工程完全指南 + +## 介绍 + +我很高兴欢迎您阅读我的最新书籍《The Art of Asking ChatGPT for High-Quality Answers: A complete Guide to Prompt Engineering Techniques》。本书是一本全面指南,介绍了各种提示技术,用于从ChatGPT中生成高质量的答案。 + +我们将探讨如何使用不同的提示工程技术来实现不同的目标。ChatGPT是一款最先进的语言模型,能够生成类似人类的文本。然而,理解如何正确地向ChatGPT提问以获得我们所需的高质量输出非常重要。而这正是本书的目的。 + +无论您是普通人、研究人员、开发人员,还是只是想在自己的领域中将ChatGPT作为个人助手的人,本书都是为您编写的。我使用简单易懂的语言,提供实用的解释,并在每个提示技术中提供了示例和提示公式。通过本书,您将学习如何使用提示工程技术来控制ChatGPT的输出,并生成符合您特定需求的文本。 + +在整本书中,我们还提供了如何结合不同的提示技术以实现更具体结果的示例。我希望您能像我写作时一样,享受阅读本书并从中获得知识。 + +
    + +## 第一章:Prompt 工程技术简介 + +什么是 Prompt 工程? + +Prompt 工程是创建提示或指导像 ChatGPT 这样的语言模型输出的过程。它允许用户控制模型的输出并生成符合其特定需求的文本。 + +ChatGPT 是一种先进的语言模型,能够生成类似于人类的文本。它建立在 Transformer 架构上,可以处理大量数据并生成高质量的文本。 + +然而,为了从 ChatGPT 中获得最佳结果,重要的是要了解如何正确地提示模型。 提示可以让用户控制模型的输出并生成相关、准确和高质量的文本。 在使用 ChatGPT 时,了解它的能力和限制非常重要。 + +该模型能够生成类似于人类的文本,但如果没有适当的指导,它可能无法始终产生期望的输出。 + +这就是 Prompt 工程的作用,通过提供清晰而具体的指令,您可以引导模型的输出并确保其相关。 + +**Prompt 公式是提示的特定格式,通常由三个主要元素组成:** + +- 任务:对提示要求模型生成的内容进行清晰而简洁的陈述。 + +- 指令:在生成文本时模型应遵循的指令。 + +- 角色:模型在生成文本时应扮演的角色。 + +在本书中,我们将探讨可用于 ChatGPT 的各种 Prompt 工程技术。我们将讨论不同类型的提示,以及如何使用它们实现您想要的特定目标。 + +
    + +## 第二章:指令提示技术 + +现在,让我们开始探索“指令提示技术”,以及如何使用它从ChatGPT中生成高质量的文本。 + +指令提示技术是通过为模型提供具体指令来引导ChatGPT的输出的一种方法。这种技术对于确保输出相关和高质量非常有用。 + +要使用指令提示技术,您需要为模型提供清晰简洁的任务,以及具体的指令以供模型遵循。 + +例如,如果您正在生成客户服务响应,您将提供任务,例如“生成响应客户查询”的指令,例如“响应应该专业且提供准确的信息”。 + +提示公式:“按照以下指示生成[任务]:[指令]” + +示例: + +**生成客户服务响应:** + +- 任务:生成响应客户查询 +- 指令:响应应该专业且提供准确的信息 +- 提示公式:“按照以下指示生成专业且准确的客户查询响应:响应应该专业且提供准确的信息。” + +**生成法律文件:** + +- 任务:生成法律文件 +- 指令:文件应符合相关法律法规 +- 提示公式:“按照以下指示生成符合相关法律法规的法律文件:文件应符合相关法律法规。” + +使用指令提示技术时,重要的是要记住指令应该清晰具体。这将有助于确保输出相关和高质量。可以将指令提示技术与下一章节中解释的“角色提示”和“种子词提示”相结合,以增强ChatGPT的输出。 + +
    + +## 第三章:角色提示 + +角色提示技术是通过为ChatGPT指定一个特定的角色来引导其输出的一种方式。这种技术对于生成针对特定上下文或受众的文本非常有用。 + +要使用角色提示技术,您需要为模型提供一个清晰具体的角色。 + +例如,如果您正在生成客户服务回复,您可以提供一个角色,如“客户服务代表”。 + +提示公式:“作为[角色]生成[任务]” + +示例: + +**生成客户服务回复:** + +- 任务:生成对客户查询的回复 +- 角色:客户服务代表 +- 提示公式:“作为客户服务代表,生成对客户查询的回复。” + +**生成法律文件:** + +- 任务:生成法律文件 +- 角色:律师 +- 提示公式:“作为律师,生成法律文件。” + +将角色提示技术与指令提示和种子词提示结合使用可以增强ChatGPT的输出。 + +**下面是一个示例,展示了如何将指令提示、角色提示和种子词提示技术结合使用:** + +- 任务:为新智能手机生成产品描述 +- 指令:描述应该是有信息量的,具有说服力,并突出智能手机的独特功能 +- 角色:市场代表 种子词:“创新的” +- 提示公式:“作为市场代表,生成一个有信息量的、有说服力的产品描述,突出新智能手机的创新功能。该智能手机具有以下功能[插入您的功能]” + +在这个示例中,指令提示用于确保产品描述具有信息量和说服力。角色提示用于确保描述是从市场代表的角度书写的。而种子词提示则用于确保描述侧重于智能手机的创新功能。 + +
    + +## 第四章:标准提示 + +标准提示是一种简单的方法,通过为模型提供一个特定的任务来引导ChatGPT的输出。例如,如果您想生成一篇新闻文章的摘要,您可以提供一个任务,如“总结这篇新闻文章”。 + +提示公式:“生成一个[任务]” + +例如: + +**生成新闻文章的摘要:** + +- 任务:总结这篇新闻文章 +- 提示公式:“生成这篇新闻文章的摘要” + +**生成一篇产品评论:** + +- 任务:为一款新智能手机撰写评论 +- 提示公式:“生成这款新智能手机的评论” + +此外,标准提示可以与其他技术(如角色提示和种子词提示)结合使用,以增强ChatGPT的输出。 + +**以下是如何将标准提示、角色提示和种子词提示技术结合使用的示例:** + +- 任务:为一台新笔记本电脑撰写产品评论 +- 说明:评论应客观、信息丰富,强调笔记本电脑的独特特点 +- 角色:技术专家 +- 种子词:“强大的” +- 提示公式:“作为一名技术专家,生成一个客观而且信息丰富的产品评论,强调新笔记本电脑的强大特点。” + +在这个示例中,标准提示技术用于确保模型生成产品评论。角色提示用于确保评论是从技术专家的角度写的。而种子词提示用于确保评论侧重于笔记本电脑的强大特点。 + +
    + +## 第五章:零、一和少样本提示 + +零样本、一样本和少样本提示是用于从ChatGPT生成文本的技术,最少或没有任何示例。当特定任务的数据有限或任务是新的且未定义时,这些技术非常有用。 + +当任务没有可用的示例时,使用零样本提示技术。模型提供一个通用任务,根据对任务的理解生成文本。 + +当任务只有一个示例可用时,使用一样本提示技术。模型提供示例,并根据对示例的理解生成文本。 + +当任务只有有限数量的示例可用时,使用少样本提示技术。模型提供示例,并根据对示例的理解生成文本。 + +提示公式:“基于[数量]个示例生成文本” + +例如: + +**为没有可用示例的新产品编写产品描述:** + +- 任务:为新的智能手表编写产品描述 + +- 提示公式:“基于零个示例为这款新智能手表生成产品描述” + +**使用一个示例生成产品比较:** + +- 任务:将新款智能手机与最新的iPhone进行比较 + +- 提示公式:“使用一个示例(最新的iPhone)为这款新智能手机生成产品比较” + +**使用少量示例生成产品评论:** + +- 任务:为新的电子阅读器撰写评论 + +- 提示公式:“使用少量示例(3个其他电子阅读器)为这款新电子阅读器生成评论” + + +这些技术可用于根据模型对任务或提供的示例的理解生成文本。 + +
    + +## 第六章:“让我们思考一下”提示 + +“让我们思考一下”提示是一种技巧,可鼓励ChatGPT生成反思和思考性的文本。这种技术适用于撰写论文、诗歌或创意写作等任务。 + +“让我们思考一下”提示的公式非常简单,即“让我们思考一下”后跟一个主题或问题。 + +例如: + +**生成一篇反思性论文:** + +- 任务:就个人成长主题写一篇反思性论文 + +- 提示公式:“让我们思考一下:个人成长” + +**生成一首诗:** + +- 任务:写一首关于季节变化的诗 + +- 提示公式:“让我们思考一下:季节变化” + + +这个提示要求对特定主题或想法展开对话或讨论。发言者邀请ChatGPT参与讨论相关主题。 + +模型提供了一个提示,作为对话或文本生成的起点。 + +然后,模型使用其训练数据和算法生成与提示相关的响应。这种技术允许ChatGPT根据提供的提示生成上下文适当且连贯的文本。 + +**要使用“让我们思考一下提示”技术与ChatGPT,您可以遵循以下步骤:** + +1. 确定您要讨论的主题或想法。 + +2. 制定一个明确表达主题或想法的提示,并开始对话或文本生成。 + +3. 用“让我们思考”或“让我们讨论”开头的提示,表明您正在启动对话或讨论。 + +**以下是使用此技术的一些提示示例:** + +- 提示:“让我们思考气候变化对农业的影响” + +- 提示:“让我们讨论人工智能的当前状态” + +- 提示:“让我们谈谈远程工作的好处和缺点” 您还可以添加开放式问题、陈述或一段您希望模型继续或扩展的文本。 + + +提供提示后,模型将使用其训练数据和算法生成与提示相关的响应,并以连贯的方式继续对话。 + +这种独特的提示有助于ChatGPT以不同的视角和角度给出答案,从而产生更具动态性和信息性的段落。 + +使用提示的步骤简单易行,可以真正提高您的写作水平。尝试一下,看看效果如何吧。 + +
    + +## 第七章:自洽提示 + +自洽提示是一种技术,用于确保ChatGPT的输出与提供的输入一致。这种技术对于事实核查、数据验证或文本生成中的一致性检查等任务非常有用。 + +自洽提示的提示公式是输入文本后跟着指令“请确保以下文本是自洽的”。 + +或者,可以提示模型生成与提供的输入一致的文本。 + +提示示例及其公式: + +**示例1:文本生成** + +- 任务:生成产品评论 + +- 指令:评论应与输入中提供的产品信息一致 + +- 提示公式:“生成与以下产品信息一致的产品评论[插入产品信息]” + +**示例2:文本摘要** + +- 任务:概括一篇新闻文章 + +- 指令:摘要应与文章中提供的信息一致 + +- 提示公式:“用与提供的信息一致的方式概括以下新闻文章[插入新闻文章]” + +**示例3:文本完成** + +- 任务:完成一个句子 + +- 指令:完成应与输入中提供的上下文一致 + +- 提示公式:“以与提供的上下文一致的方式完成以下句子[插入句子]” + +**示例4:** + +1. **事实核查:** + + 任务:检查给定新闻文章的一致性 + + 输入文本:“文章中陈述该城市的人口为500万,但后来又说该城市的人口为700万。” + + 提示公式:“请确保以下文本是自洽的:文章中陈述该城市的人口为500万,但后来又说该城市的人口为700万。” + +2. **数据验证:** + + 任务:检查给定数据集的一致性 + + 输入文本:“数据显示7月份的平均温度为30度,但最低温度记录为20度。” + + 提示公式:“请确保以下文本是自洽的:数据显示7月份的平均温度为30度,但最低温度记录为20度。” + +
    + +## 第八章:种子词提示 + +种子词提示是一种通过提供特定的种子词或短语来控制ChatGPT输出的技术。种子词提示的提示公式是种子词或短语,后跟指令“请根据以下种子词生成文本”。 + +示例: + +**文本生成:** + +- 任务:编写一篇有关龙的故事 +- 种子词:“龙” +- 提示公式:“请根据以下种子词生成文本:龙” + +**语言翻译:** + +- 任务:将一句话从英语翻译成西班牙语 +- 种子词:“你好” +- 提示公式:“请根据以下种子词生成文本:你好” + +这种技术允许模型生成与种子词相关的文本并对其进行扩展。这是一种控制模型生成文本与某个特定主题或背景相关的方式。 + +种子词提示可以与角色提示和指令提示相结合,以创建更具体和有针对性的生成文本。通过提供种子词或短语,模型可以生成与该种子词或短语相关的文本,并通过提供有关期望输出和角色的信息,模型可以以特定于角色或指令的风格或语气生成文本。这样可以更好地控制生成的文本,并可用于各种应用程序。 + +以下是提示示例及其公式: + +**示例1:文本生成** + +- 任务:编写一首诗 +- 指令:诗应与种子词“爱”相关,并以十四行诗的形式书写。 +- 角色:诗人 +- 提示公式:“作为诗人,根据以下种子词生成与“爱”相关的十四行诗:” + +**示例2:文本完成** + +- 任务:完成一句话 +- 指令:完成应与种子词“科学”相关,并以研究论文的形式书写。 +- 角色:研究员 +- 提示公式:“作为研究员,请在与种子词“科学”相关且以研究论文的形式书写的情况下完成以下句子:[插入句子]” + +**示例3:文本摘要** + +- 任务:摘要一篇新闻文章 +- 指令:摘要应与种子词“政治”相关,并以中立和公正的语气书写。 +- 角色:记者 +- 提示公式:“作为记者,请以中立和公正的语气摘要以下新闻文章,与种子词“政治”相关:[插入新闻文章]” + +
    + +## 第九章:知识生成提示 + +知识生成提示是一种从ChatGPT中引出新的、原创的信息的技术。 + +知识生成提示的公式是“请生成关于X的新的和原创的信息”,其中X是感兴趣的主题。 + +这是一种利用模型预先存在的知识来生成新的信息或回答问题的技术。 + +要将此提示与ChatGPT一起使用,需要将问题或主题作为输入提供给模型,以及指定所生成文本的任务或目标的提示。 + +提示应包括有关所需输出的信息,例如要生成的文本类型以及任何特定的要求或限制。 + +以下是提示示例及其公式: + +**示例1:知识生成** + +- 任务:生成有关特定主题的新信息 +- 说明:生成的信息应准确且与主题相关 +- 提示公式:“生成有关[特定主题]的新的准确信息” + +**示例2:问答** + +- 任务:回答问题 +- 说明:答案应准确且与问题相关 +- 提示公式:“回答以下问题:[插入问题]” + +**示例3:知识整合** + +- 任务:将新信息与现有知识整合 +- 说明:整合应准确且与主题相关 +- 提示公式:“将以下信息与有关[特定主题]的现有知识整合:[插入新信息]” + +**示例4:数据分析** + +- 任务:从给定的数据集中生成有关客户行为的见解 +- 提示公式:“请从这个数据集中生成有关客户行为的新的和原创的信息” + +
    + +## 第十章:知识整合提示 + +这种技术利用模型的现有知识来整合新信息或连接不同的信息片段。 + +这种技术对于将现有知识与新信息相结合,以生成更全面的特定主题的理解非常有用。 + +**如何与ChatGPT一起使用:** + +- 模型应该提供新信息和现有知识作为输入,以及指定生成文本的任务或目标的提示。 +- 提示应包括有关所需输出的信息,例如要生成的文本类型以及任何特定的要求或限制。 + +提示示例及其公式: + +**示例1:知识整合** + +- 任务:将新信息与现有知识整合 +- 说明:整合应准确且与主题相关 +- 提示公式:“将以下信息与关于[具体主题]的现有知识整合:[插入新信息]” + +**示例2:连接信息片段** + +- 任务:连接不同的信息片段 +- 说明:连接应相关且逻辑清晰 +- 提示公式:“以相关且逻辑清晰的方式连接以下信息片段:[插入信息1] [插入信息2]” + +**示例3:更新现有知识** + +- 任务:使用新信息更新现有知识 +- 说明:更新的信息应准确且相关 +- 提示公式:“使用以下信息更新[具体主题]的现有知识:[插入新信息]” + +
    + +## 第十一章:多项选择提示 + +这种技术向模型提供一个问题或任务以及一组预定义的选项作为潜在答案。 + +该技术对于生成仅限于特定选项集的文本非常有用,可用于问答、文本完成和其他任务。模型可以生成仅限于预定义选项的文本。 + +要使用ChatGPT的多项选择提示,需要向模型提供一个问题或任务作为输入,以及一组预定义的选项作为潜在答案。提示还应包括有关所需输出的信息,例如要生成的文本类型以及任何特定要求或限制。 + +提示示例及其公式: + +**示例1:问答** + +- 任务:回答一个多项选择题 +- 说明:答案应该是预定义的选项之一 +- 提示公式:“通过选择以下选项之一回答以下问题:[插入问题] [插入选项1] [插入选项2] [插入选项3]” + +**示例2:文本完成** + +- 任务:使用预定义选项之一完成句子 +- 说明:完成应该是预定义的选项之一 +- 提示公式:“通过选择以下选项之一完成以下句子:[插入句子] [插入选项1] [插入选项2] [插入选项3]” + +**示例3:情感分析** + +- 任务:将文本分类为积极、中立或消极 +- 说明:分类应该是预定义的选项之一 +- 提示公式:“通过选择以下选项之一,将以下文本分类为积极、中立或消极:[插入文本] [积极] [中立] [消极]” + +
    + +## 第十二章:可解释的软提示 + +可解释的软提示是一种技术,可以在提供一定的灵活性的同时控制模型生成的文本。它通过提供一组受控输入和关于所需输出的附加信息来实现。这种技术可以生成更具解释性和可控性的生成文本。 + +提示示例及其公式: + +**示例1:文本生成** + +- 任务:生成一个故事 +- 指令:故事应基于一组给定的角色和特定的主题 +- 提示公式:“基于以下角色生成故事:[插入角色]和主题:[插入主题]” + +**示例2:文本完成** + +- 任务:完成一句话 +- 指令:完成应以特定作者的风格为基础 +- 提示公式:“以[特定作者]的风格完成以下句子:[插入句子]” + +**示例3:语言建模** + +- 任务:以特定风格生成文本 +- 指令:文本应以特定时期的风格为基础 +- 提示公式:“以[特定时期]的风格生成文本:[插入上下文]” + +
    + +## 第十三章:控制生成提示 + +控制生成提示是一种技术,可让模型在生成文本时对输出进行高度控制。 + +这可以通过提供一组特定的输入来实现,例如模板、特定词汇或一组约束条件,这些输入可用于指导生成过程。 + +以下是一些示例和它们的公式: + +**示例1:文本生成** + +- 任务:生成一个故事 +- 说明:该故事应基于特定的模板 +- 提示公式:“根据以下模板生成故事:[插入模板]” + +**示例2:文本补全** + +- 任务:完成一句话 +- 说明:完成应使用特定的词汇 +- 提示公式:“使用以下词汇完成以下句子:[插入词汇]:[插入句子]” + +**示例3:语言建模** + +- 任务:以特定风格生成文本 +- 说明:文本应遵循一组特定的语法规则 +- 提示公式:“生成遵循以下语法规则的文本:[插入规则]:[插入上下文]” + +通过提供一组特定的输入来指导生成过程,控制生成提示使得生成的文本更具可控性和可预测性。 + +
    + +## 第十四章:问答提示 + +问答提示是一种技术,可以让模型生成回答特定问题或任务的文本。通过将问题或任务与可能与问题或任务相关的任何其他信息一起作为输入提供给模型来实现此目的。 + +一些提示示例及其公式如下: + +**示例1:事实问题回答** + +- 任务:回答一个事实性问题 +- 说明:答案应准确且相关 +- 提示公式:“回答以下事实问题:[插入问题]” + +**示例2:定义** + +- 任务:提供一个词的定义 +- 说明:定义应准确 +- 提示公式:“定义以下词汇:[插入单词]” + +**示例3:信息检索** + +- 任务:从特定来源检索信息 +- 说明:检索到的信息应相关 +- 提示公式:“从以下来源检索有关[特定主题]的信息:[插入来源]” 这对于问答和信息检索等任务非常有用。 + +
    + +## 第十五章:概述提示 + +概述提示是一种技术,允许模型在保留其主要思想和信息的同时生成给定文本的较短版本。 + +这可以通过将较长的文本作为输入提供给模型并要求其生成该文本的摘要来实现。 + +这种技术对于文本概述和信息压缩等任务非常有用。 + +**如何在ChatGPT中使用:** + +- 应该向模型提供较长的文本作为输入,并要求其生成该文本的摘要。 +- 提示还应包括有关所需输出的信息,例如摘要的所需长度和任何特定要求或限制。 + +提示示例及其公式: + +**示例1:文章概述** + +- 任务:概述新闻文章 +- 说明:摘要应是文章主要观点的简要概述 +- 提示公式:“用一句简短的话概括以下新闻文章:[插入文章]” + +**示例2:会议记录** + +- 任务:概括会议记录 +- 说明:摘要应突出会议的主要决策和行动 +- 提示公式:“通过列出主要决策和行动来总结以下会议记录:[插入记录]” + +**示例3:书籍摘要** + +- 任务:总结一本书 +- 说明:摘要应是书的主要观点的简要概述 +- 提示公式:“用一段简短的段落总结以下书籍:[插入书名]” + +
    + +## 第十六章:对话提示 + +对话提示是一种技术,允许模型生成模拟两个或更多实体之间对话的文本。通过为模型提供一个上下文和一组角色或实体,以及它们的角色和背景,并要求模型在它们之间生成对话。 + +因此,应为模型提供上下文和一组角色或实体,以及它们的角色和背景。还应向模型提供有关所需输出的信息,例如对话或交谈的类型以及任何特定的要求或限制。 + +提示示例及其公式: + +**示例1:对话生成** + +- 任务:生成两个角色之间的对话 +- 说明:对话应自然且与给定上下文相关 +- 提示公式:“在以下情境中生成以下角色之间的对话[插入角色]” + +**示例2:故事写作** + +- 任务:在故事中生成对话 +- 说明:对话应与故事的角色和事件一致 +- 提示公式:“在以下故事中生成以下角色之间的对话[插入故事]” + +**示例3:聊天机器人开发** + +- 任务:为客服聊天机器人生成对话 +- 说明:对话应专业且提供准确的信息 +- 提示公式:“在客户询问[插入主题]时,为客服聊天机器人生成专业和准确的对话” + +因此,这种技术对于对话生成、故事写作和聊天机器人开发等任务非常有用。 + +
    + +## 第十七章:对抗性提示 + +对抗性提示是一种技术,它允许模型生成抵抗某些类型的攻击或偏见的文本。这种技术可用于训练更为稳健和抵抗某些类型攻击或偏见的模型。 + +要在ChatGPT中使用对抗性提示,需要为模型提供一个提示,该提示旨在使模型难以生成符合期望输出的文本。提示还应包括有关所需输出的信息,例如要生成的文本类型和任何特定要求或约束。 + +提示示例及其公式: + +**示例1:用于文本分类的对抗性提示** + +- 任务:生成被分类为特定标签的文本 +- 说明:生成的文本应难以分类为特定标签 +- 提示公式:“生成难以分类为[插入标签]的文本” + +**示例2:用于情感分析的对抗性提示** + +- 任务:生成难以分类为特定情感的文本 +- 说明:生成的文本应难以分类为特定情感 +- 提示公式:“生成难以分类为具有[插入情感]情感的文本” + +**示例3:用于语言翻译的对抗性提示** + +- 任务:生成难以翻译的文本 +- 说明:生成的文本应难以翻译为目标语言 +- 提示公式:“生成难以翻译为[插入目标语言]的文本” + +
    + +## 第十八章:聚类提示 + +聚类提示是一种技术,它可以让模型根据某些特征或特点将相似的数据点分组在一起。 + +通过提供一组数据点并要求模型根据某些特征或特点将它们分组成簇,可以实现这一目标。 + +这种技术在数据分析、机器学习和自然语言处理等任务中非常有用。 + +**如何在ChatGPT中使用:** + +应该向模型提供一组数据点,并要求它根据某些特征或特点将它们分组成簇。提示还应包括有关所需输出的信息,例如要生成的簇数和任何特定的要求或约束。 + +提示示例及其公式: + +**示例1:客户评论的聚类** + +- 任务:将相似的客户评论分组在一起 +- 说明:应根据情感将评论分组 +- 提示公式:“将以下客户评论根据情感分组成簇:[插入评论]” + +**示例2:新闻文章的聚类** + +- 任务:将相似的新闻文章分组在一起 +- 说明:应根据主题将文章分组 +- 提示公式:“将以下新闻文章根据主题分组成簇:[插入文章]” + +**示例3:科学论文的聚类** + +- 任务:将相似的科学论文分组在一起 +- 说明:应根据研究领域将论文分组 +- 提示公式:“将以下科学论文根据研究领域分组成簇:[插入论文]” + +
    + +## 第十九章:强化学习提示 + +强化学习提示是一种技术,可以使模型从过去的行动中学习,并随着时间的推移提高其性能。要在ChatGPT中使用强化学习提示,需要为模型提供一组输入和奖励,并允许其根据接收到的奖励调整其行为。提示还应包括有关期望输出的信息,例如要完成的任务以及任何特定要求或限制。这种技术对于决策制定、游戏玩法和自然语言生成等任务非常有用。 + +提示示例及其公式: + +**示例1:用于文本生成的强化学习** + +- 任务:生成与特定风格一致的文本 +- 说明:模型应根据为生成与特定风格一致的文本而接收到的奖励来调整其行为 +- 提示公式:“使用强化学习来生成与以下风格一致的文本[插入风格]” + +**示例2:用于语言翻译的强化学习** + +- 任务:将文本从一种语言翻译成另一种语言 +- 说明:模型应根据为生成准确翻译而接收到的奖励来调整其行为 +- 提示公式:“使用强化学习将以下文本[插入文本]从[插入语言]翻译成[插入语言]” + +**示例3:用于问答的强化学习** + +- 任务:回答问题 +- 说明:模型应根据为生成准确答案而接收到的奖励来调整其行为 +- 提示公式:“使用强化学习来回答以下问题[插入问题]” + +
    + +## 第二十章:课程学习提示 + +课程学习是一种技术,允许模型通过先训练简单任务,逐渐增加难度来学习复杂任务。 + +要在ChatGPT中使用课程学习提示,模型应该提供一系列任务,这些任务逐渐增加难度。 + +提示还应包括有关期望输出的信息,例如要完成的最终任务以及任何特定要求或约束条件。 + +此技术对自然语言处理、图像识别和机器学习等任务非常有用。 + +提示示例及其公式: + +**示例1:用于文本生成的课程学习** + +- 任务:生成与特定风格一致的文本 +- 说明:模型应该在移动到更复杂的风格之前先在简单的风格上进行训练。 +- 提示公式:“使用课程学习来生成与以下风格[插入风格]一致的文本,按照以下顺序[插入顺序]。” + +**示例2:用于语言翻译的课程学习** + +- 任务:将文本从一种语言翻译成另一种语言 +- 说明:模型应该在移动到更复杂的语言之前先在简单的语言上进行训练。 +- 提示公式:“使用课程学习将以下语言[插入语言]的文本翻译成以下顺序[插入顺序]。” + +**示例3:用于问题回答的课程学习** + +- 任务:回答问题 +- 说明:模型应该在移动到更复杂的问题之前先在简单的问题上进行训练。 +- 提示公式:“使用课程学习来回答以下问题[插入问题],按照以下顺序[插入顺序]生成答案。” + +
    + +## 第二十一章:情感分析提示 + +情感分析是一种技术,允许模型确定文本的情绪色彩或态度,例如它是积极的、消极的还是中立的。 + +要在ChatGPT中使用情感分析提示,模型应该提供一段文本并要求根据其情感分类。 + +提示还应包括关于所需输出的信息,例如要检测的情感类型(例如积极的、消极的、中立的)和任何特定要求或约束条件。 + +提示示例及其公式: + +**示例1:客户评论的情感分析** + +- 任务:确定客户评论的情感 +- 说明:模型应该将评论分类为积极的、消极的或中立的 +- 提示公式:“对以下客户评论进行情感分析[插入评论],并将它们分类为积极的、消极的或中立的。” + +**示例2:推文的情感分析** + +- 任务:确定推文的情感 +- 说明:模型应该将推文分类为积极的、消极的或中立的 +- 提示公式:“对以下推文进行情感分析[插入推文],并将它们分类为积极的、消极的或中立的。” + +**示例3:产品评论的情感分析** + +- 任务:确定产品评论的情感 +- 说明:模型应该将评论分类为积极的、消极的或中立的 +- 提示公式:“对以下产品评论进行情感分析[插入评论],并将它们分类为积极的、消极的或中立的。” + +这种技术对自然语言处理、客户服务和市场研究等任务非常有用。 + +
    + +## 第二十二章:命名实体识别提示 + +命名实体识别(NER)是一种技术,它可以使模型识别和分类文本中的命名实体,例如人名、组织机构、地点和日期等。 + +要在ChatGPT中使用命名实体识别提示,需要向模型提供一段文本,并要求它识别和分类文本中的命名实体。 + +提示还应包括有关所需输出的信息,例如要识别的命名实体类型(例如人名、组织机构、地点、日期)以及任何特定要求或约束条件。 + +提示示例及其公式: + +**示例1:新闻文章中的命名实体识别** + +- 任务:在新闻文章中识别和分类命名实体 +- 说明:模型应识别和分类人名、组织机构、地点和日期 +- 提示公式:“在以下新闻文章[插入文章]上执行命名实体识别,并识别和分类人名、组织机构、地点和日期。” + +**示例2:法律文件中的命名实体识别** + +- 任务:在法律文件中识别和分类命名实体 +- 说明:模型应识别和分类人名、组织机构、地点和日期 +- 提示公式:“在以下法律文件[插入文件]上执行命名实体识别,并识别和分类人名、组织机构、地点和日期。” + +**示例3:研究论文中的命名实体识别** + +- 任务:在研究论文中识别和分类命名实体 +- 说明:模型应识别和分类人名、组织机构、地点和日期 +- 提示公式:“在以下研究论文[插入论文]上执行命名实体识别,并识别和分类人名、组织机构、地点和日期。” + +
    + +## 第二十三章:文本分类提示 + +文本分类是一种技术,它可以让模型将文本分成不同的类别。这种技术对于自然语言处理、文本分析和情感分析等任务非常有用。 + +需要注意的是,文本分类和情感分析是不同的。情感分析特别关注于确定文本中表达的情感或情绪。这可能包括确定文本表达了积极、消极还是中性的情感。情感分析通常用于客户评论、社交媒体帖子和其他需要表达情感的文本。 + +要在ChatGPT中使用文本分类提示,模型需要提供一段文本,并要求它根据预定义的类别或标签进行分类。提示还应包括有关所需输出的信息,例如类别或标签的数量以及任何特定的要求或约束。 + +提示示例及其公式: + +**示例1:对客户评论进行文本分类** + +- 任务:将客户评论分类为不同的类别,例如电子产品、服装和家具 +- 说明:模型应根据评论的内容对其进行分类 +- 提示公式:“对以下客户评论 [插入评论] 进行文本分类,并根据其内容将其分类为不同的类别,例如电子产品、服装和家具。” + +**示例2:对新闻文章进行文本分类** + +- 任务:将新闻文章分类为不同的类别,例如体育、政治和娱乐 +- 说明:模型应根据文章的内容对其进行分类 +- 提示公式:“对以下新闻文章 [插入文章] 进行文本分类,并根据其内容将其分类为不同的类别,例如体育、政治和娱乐。” + +**示例3:对电子邮件进行文本分类** + +- 任务:将电子邮件分类为不同的类别,例如垃圾邮件、重要邮件或紧急邮件 +- 说明:模型应根据电子邮件的内容和发件人对其进行分类 +- 提示公式:“对以下电子邮件 [插入电子邮件] 进行文本分类,并根据其内容和发件人将其分类为不同的类别,例如垃圾邮件、重要邮件或紧急邮件。” + +
    + +## 第二十四章:文本生成提示 + +文本生成提示与本书中提到的其他提示技术相关,例如:零、一、几次提示,受控生成提示,翻译提示,语言建模提示,句子补全提示等。这些提示都与生成文本有关,但它们在生成文本的方式和放置在生成文本上的特定要求或限制方面有所不同。文本生成提示可用于微调预训练模型或训练新模型以执行特定任务。 + +提示示例及其公式: + +**示例1:故事创作的文本生成** + +- 任务:根据给定的提示生成故事 +- 说明:故事应至少包含1000个单词,并包括一组特定的角色和情节。 +- 提示公式:“根据以下提示[插入提示]生成一个至少包含1000个单词,包括角色[插入角色]和情节[插入情节]的故事。” + +**示例2:语言翻译的文本生成** + +- 任务:将给定的文本翻译成另一种语言 +- 说明:翻译应准确并符合习惯用语。 +- 提示公式:“将以下文本[插入文本]翻译成[插入目标语言],并确保其准确且符合习惯用语。” + +**示例3:文本完成的文本生成** + +- 任务:完成给定的文本 +- 说明:生成的文本应与输入文本连贯一致。 +- 提示公式:“完成以下文本[插入文本],并确保其连贯一致且符合输入文本。” + +
    + +## 结语 + +正如本书中所探讨的那样,快速工程是一种利用像ChatGPT这样的语言模型获得高质量答案的强大工具。通过精心设计各种技巧的提示,我们可以引导模型生成符合我们特定需求和要求的文本。 + +在第二章中,我们讨论了如何使用指令提示向模型提供清晰明确的指导。在第三章中,我们探讨了如何使用角色提示生成特定的语音或风格的文本。在第四章中,我们研究了如何使用标准提示作为微调模型性能的起点。我们还研究了几种高级提示技术,例如Zero、One和Few Shot Prompting、Self-Consistency、Seed-word Prompt、Knowledge Generation Prompt、Knowledge Integration prompts、Multiple Choice prompts、Interpretable Soft Prompts、Controlled generation prompts、Question-answering prompts、Summarization prompts、Dialogue prompts、Adversarial prompts、Clustering prompts、Reinforcement learning prompts、Curriculum learning prompts、Sentiment analysis prompts、Named entity recognition prompts和Text classification prompts(对应章节的名字)。 + +这些技术中的每一种都可以以不同的方式使用,以实现各种不同的结果。随着您继续使用ChatGPT和其他语言模型,值得尝试不同的技巧组合,以找到最适合您特定用例的方法。 + +最后,您可以查看我写的其他主题的书籍。 + +感谢您阅读整本书。期待在我的其他书中与您见面。 + +(本文翻译自《The Art of Asking ChatGPT for High-Quality Answers A Complete Guide to Prompt Engineering Techniques》这本书,本文的翻译全部由ChatGpt完成,我只是把翻译内容给稍微排版了一下。做完了才发现这个工作早就有人做过了...下面是我以此事件让New Bing编写的一个小故事,希望大家喜欢) + +> 他终于画完了他的画,心满意足地把它挂在了墙上。他觉得这是他一生中最伟大的作品,无人能及。他邀请了所有的朋友来欣赏,期待着他们的赞美和惊叹。 可是当他们看到画时,却没有一个人说话。他们只是互相对视,然后低头咳嗽,或者假装看手机。他感到很奇怪,难道他们都不懂艺术吗?难道他们都没有眼光吗? 他忍不住问其中一个朋友:“你觉得我的画怎么样?” 朋友犹豫了一下,说:“嗯……其实……这个画……我以前在哪里见过。” “见过?你在哪里见过?”他惊讶地问。 “就在……就在那边啊。”朋友指了指墙角的一个小框架,“那不就是你上个月买回来的那幅名画吗?你怎么把它照抄了一遍? ——New Bing + [这就是那幅名画]: http://yesaiwen.com/art-of-asking-chatgpt-for-high-quality-answ-engineering-techniques/#i-3 "《如何向ChatGPT提问并获得高质量的答案》" \ No newline at end of file diff --git a/knowledge_base/samples/isssues_merge/langchain-ChatGLM_closed.csv b/knowledge_base/samples/isssues_merge/langchain-ChatGLM_closed.csv index 9af3a6c0..b1ca2ac9 100644 --- a/knowledge_base/samples/isssues_merge/langchain-ChatGLM_closed.csv +++ b/knowledge_base/samples/isssues_merge/langchain-ChatGLM_closed.csv @@ -1,173 +1,173 @@ -,title,file,url,detail,id -0,加油~以及一些建议,2023-03-31.0002,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/2,加油,我认为你的方向是对的。,0 -1,当前的运行环境是什么,windows还是Linux,2023-04-01.0003,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/3,当前的运行环境是什么,windows还是Linux,python是什么版本?,1 -2,请问这是在CLM基础上运行吗?,2023-04-01.0004,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/4,请问是不是需要本地安装好clm并正常运行的情况下,再按文中的步骤执行才能运行起来?,2 -3,[复现问题] 构造 prompt 时从知识库中提取的文字乱码,2023-04-01.0005,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/5,hi,我在尝试复现 README 中的效果,也使用了 ChatGLM-6B 的 README 作为输入文本,但发现从知识库中提取的文字是乱码,导致构造的 prompt 不可用。想了解如何解决这个问题。,3 -4,后面能否加入上下文对话功能?,2023-04-02.0006,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/6,目前的get_wiki_agent_answer函数中已经实现了历史消息传递的功能,后面我再确认一下是否有langchain中model调用过程中是否传递了chat_history。,4 -5,请问:纯cpu可以吗?,2023-04-03.0007,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/7,很酷的实现,极大地开拓了我的眼界!很顺利的在gpu机器上运行了,5 -6,运行报错:AttributeError: 'NoneType' object has no attribute 'message_types_by_name',2023-04-03.0008,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/8,报错:,6 -7,运行环境:GPU需要多大的?,2023-04-03.0009,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/9,如果按照THUDM/ChatGLM-6B的说法,使用的GPU大小应该在13GB左右,但运行脚本后,占用了24GB还不够。,7 -8,请问本地知识的格式是什么?,2023-04-03.0010,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/10,已测试格式包括docx、md文件中的文本信息,具体格式可以参考 [langchain文档](https://python.langchain.com/en/latest/modules/indexes/document_loaders/examples/unstructured_file.html?highlight=pdf#),8 -9,24G的显存还是爆掉了,是否支持双卡运行,2023-04-03.0011,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/11,RuntimeError: CUDA out of memory. Tried to allocate 96.00 MiB (GPU 0; 23.70 GiB total capacity; 22.18 GiB already allocated; 12.75 MiB free; 22.18 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF,9 -10,你怎么知道embeddings方式和模型训练时候的方式是一样的?,2023-04-03.0012,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/12,embedding和LLM的方式不用一致,embedding能够解决语义检索的需求就行。这个项目里用到embedding是在对本地知识建立索引和对问句转换成向量的过程。,10 -11,是否能提供本地知识文件的格式?,2023-04-04.0013,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/13,是否能提供本地知识文件的格式?,11 -12,是否可以像清华原版跑在8G一以下的卡?,2023-04-04.0016,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/16,是否可以像清华原版跑在8G一以下的卡?我的8G卡爆显存了🤣🤣🤣,12 -13,请教一下langchain协调使用向量库和chatGLM工作的,2023-04-05.0018,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/18,代码里面这段是创建问答模型的,会接入ChatGLM和本地语料的向量库,langchain回答的时候是怎么个优先顺序?先搜向量库,没有再找chatglm么? 还是什么机制?,13 -14,在mac m2max上抛出了ValueError: 150001 is not in list这个异常,2023-04-05.0019,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/19,我把chatglm_llm.py加载模型的代码改成如下,14 -15,程序运行后一直卡住,2023-04-05.0020,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/20,感谢作者的付出,不过本人在运行时出现了问题,请大家帮助。,15 -16,问一下chat_history的逻辑,2023-04-06.0022,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/22,感谢开源。,16 -17,为什么每次运行都会loading checkpoint,2023-04-06.0023,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/23,我把这个embeding模型下载到本地后,无法正常启动。,17 -18,本地知识文件能否上传一些示例?,2023-04-06.0025,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/25,如题,怎么构造知识文件,效果更好?能否提供一个样例,18 -19,What version of you are using?,2023-04-06.0026,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/26,"Hi Panda, I saw the `pip install -r requirements` command in README, and want to confirm you are using python2 or python3? because my pip and pip3 version are all is 22.3.",19 -20,有兴趣交流本项目应用的朋友可以加一下微信群,2023-04-07.0027,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/27,![IMG_1630](https://user-images.githubusercontent.com/5668498/230533162-8b9bfcdd-249c-4efe-b066-4f9ba2ce9f23.jpeg),20 -21,本地知识越多,回答时检索的时间是否会越长,2023-04-07.0029,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/29,是的 因为需要进行向量匹配检索,21 -22,爲啥最後還是報錯 哭。。,2023-04-07.0030,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/30,Failed to import transformers.models.t5.configuration_t5 because of the following error (look up to see,22 -23,对话到第二次的时候就报错UnicodeDecodeError: 'utf-8' codec can't decode,2023-04-07.0031,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/31,对话第一次是没问题的,模型返回输出后又给到请输入你的问题,我再输入问题就报错,23 -24,用的in4的量化版本,推理的时候显示需要申请10Gb的显存,2023-04-07.0033,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/33,"File ""/root/.cache/huggingface/modules/transformers_modules/chatglm-6b-int4-qe/modeling_chatglm.py"", line 581, in forward",24 -25,使用colab运行,python3.9,提示包导入有问题,2023-04-07.0034,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/34,"from ._util import is_directory, is_path",25 -26,运行失败,Loading checkpoint未达到100%被kill了,请问下是什么原因?,2023-04-07.0035,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/35,日志如下:,26 -27,弄了个交流群,自己弄好多细节不会,大家技术讨论 加connection-image 我来拉你,2023-04-08.0036,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/36,自己搞好多不清楚的,一起来弄吧。。准备搞个部署问题的解决文档出来,27 -28,Error using the new version with langchain,2023-04-09.0043,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/43,Error with the new changes:,28 -29,程序报错torch.cuda.OutOfMemoryError如何解决?,2023-04-10.0044,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/44,报错详细信息如下:,29 -30,qa的训练数据格式是如何设置的,2023-04-10.0045,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/45,本项目不是使用微调的方式,所以并不涉及到训练过程。,30 -31,The FileType.UNK file type is not supported in partition. 解决办法,2023-04-10.0046,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/46,ValueError: Invalid file /home/yawu/Documents/langchain-ChatGLM-master/data. The FileType.UNK file type is not supported in partition.,31 -32,如何读取多个txt文档?,2023-04-10.0047,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/47,如题,请教一下如何读取多个txt文档?示例代码中只给了读一个文档的案例,这个input我换成string之后也只能指定一个文档,无法用通配符指定多个文档,也无法传入多个文件路径的列表。,32 -33,nltk package unable to either download or load local nltk_data folder,2023-04-10.0049,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/49,I'm running this project on an offline Windows Server environment so I download the Punkt and averaged_perceptron_tagger tokenizer in this directory:,33 -34,requirements.txt中需要指定langchain版本,2023-04-11.0055,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/55,langchain版本0.116下无法引入RetrievalQA,需要指定更高版本(0.136版本下无问题),34 -35,Demo演示无法给出输出内容,2023-04-12.0059,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/59,你好,测试了项目自带新闻稿示例和自行上传的一个文本,可以加载进去,但是无法给出答案,请问属于什么情况,如何解决,谢谢。PS: 1、今天早上刚下载全部代码;2、硬件服务器满足要求;3、按操作说明正常操作。,35 -36,群人数过多无法进群,求帮忙拉进群,2023-04-12.0061,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/61,您好,您的群人数超过了200人,目前无法通过二维码加群,请问您方便加我微信拉我进群吗?万分感谢,36 -37,群人数已满,求大佬拉入群,2023-04-12.0062,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/62,已在README中更新拉群二维码,37 -38,requirements中langchain版本错误,2023-04-12.0065,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/65,langchain版本应该是0.0.12而不是0.0.120,38 -39,Linux : Searchd in,2023-04-13.0068,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/68,import nltk,39 -40,No sentence-transformers model found,2023-04-13.0069,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/69,加载不了这个模型,错误原因是找不到这个模型,但是路径是配置好了的,40 -41,Error loading punkt: ",58 -59,为啥放到方法调用会出错,这个怎么处理?,2023-04-20.0150,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/150,```python,59 -60,No sentence-transformers model found with name C:\Users\Administrator/.cache\torch\sentence_transformers\GanymedeNil_text2vec-large-chinese. Creating a new one with MEAN pooling.,2023-04-21.0154,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/154,卡在这块很久是正常现象吗,60 -61,微信群需要邀请才能加入,2023-04-21.0155,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/155,RT,给个个人联系方式白,61 -62,No sentence-transformers model found with name GanymedeNil/text2vec-large-chinese. Creating a new one with MEAN pooling,2023-04-21.0156,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/156,ls GanymedeNil/text2vec-large-chinese,62 -63,embedding会加载两次,2023-04-23.0159,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/159,你好,为什么要这样设置呢,这样会加载两次呀。,63 -64,扫二维码加的那个群,群成员满了进不去了,2023-04-23.0160,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/160,如题,64 -65,执行python3 cli_demo.py 报错AttributeError: 'NoneType' object has no attribute 'chat',2023-04-24.0163,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/163,"刚开始怀疑是内存不足问题,换成int4,int4-qe也不行,有人知道是什么原因吗",65 -66,匹配得分,2023-04-24.0167,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/167,在示例cli_demo.py中返回的匹配文本没有对应的score,可以加上这个feature吗,66 -67,大佬有计划往web_ui.py加入打字机功能吗,2023-04-25.0170,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/170,目前在载入了知识库后,单张V100 32G在回答垂直领域的问题时也需要20S以上,没有打字机逐字输出的使用体验还是比较煎熬的....,67 -68,Is it possible to use a verctorDB for the embedings?,2023-04-25.0171,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/171,"when I play, I have to load the local data again and again when to start. I wonder if it is possible to use",68 -69,请问通过lora训练官方模型得到的微调模型文件该如何加载?,2023-04-25.0173,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/173,通过lora训练的方式得到以下文件:,69 -70,from langchain.chains import RetrievalQA的代码在哪里?,2023-04-25.0174,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/174,local_doc_qa.py,70 -71,哪里有knowledge_based_chatglm.py文件?怎么找不到了??是被替换成cli_demo.py文件了吗?,2023-04-26.0175,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/175,哪里有knowledge_based_chatglm.py文件?怎么找不到了??是被替换成cli_demo.py文件了吗?,71 -72,AttributeError: 'Chatbot' object has no attribute 'value',2023-04-26.0177,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/177,Traceback (most recent call last):,72 -73,控制台调api.py报警告,2023-04-26.0178,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/178,"you must pass the application as an import string to enable ""reload"" or ""workers""",73 -74,如何加入群聊,2023-04-27.0183,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/183,微信群超过200人了,需要邀请,如何加入呢?,74 -75,如何将Chatglm和本地知识相结合,2023-04-27.0185,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/185,您好,我想请教一下怎么才能让知识库匹配到的文本和chatglm生成的相结合,而不是说如果没搜索到,就说根据已知信息无法回答该问题,谢谢,75 -76,一点建议,2023-04-27.0189,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/189,1.weiui的get_vector_store方法里面添加一个判断以兼容gradio版本导致的上传异常,76 -77,windows环境下,按照教程,配置好conda环境,git完项目,修改完模型路径相关内容后,运行demo报错缺少,2023-04-28.0194,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/194,报错代码如下:,77 -78,ValueError: too many values to unpack (expected 2),2023-04-28.0198,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/198,"When i tried to use the non-streaming, `ValueError: too many values to unpack (expected 2)` error came out.",78 -79,加载doc后覆盖原本知识,2023-04-28.0201,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/201,加载较大量级的私有知识库后,原本的知识会被覆盖,79 -80,自定义知识库回答效果很差,2023-04-28.0203,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/203,"请问加了自定义知识库知识库,回答效果很差,是因为数据量太小的原因么",80 -81,python310下,安装pycocotools失败,提示低版本cython,实际已安装高版本,2023-04-29.0208,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/208,RT,纯离线环境安装,依赖安装的十分艰难,最后碰到pycocotools,始终无法安装上,求教方法!,81 -82,[FEATURE] 支持 RWKV 模型(目前已有 pip package & rwkv.cpp 等等),2023-05-01.0216,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/216,您好,我是 RWKV 的作者,介绍见:https://zhuanlan.zhihu.com/p/626083366,82 -83,[BUG] 为啥主机/服务器不联网不能正常启动服务?,2023-05-02.0220,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/220,**问题描述 / Problem Description**,83 -84,[BUG] 简洁阐述问题 / Concise description of the issue,2023-05-03.0222,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/222,**local variable 'torch' referenced before assignment**,84 -85,不支持txt文件的中文输入,2023-05-04.0235,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/235,"vs_path, _ = local_doc_qa.init_knowledge_vector_store(filepath)",85 -86,文件均未成功加载,请检查依赖包或替换为其他文件再次上传。 文件未成功加载,请重新上传文件,2023-05-05.0237,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/237,请大佬帮忙解决,谢谢!,86 -87,[BUG] 使用多卡时chatglm模型加载两次,2023-05-05.0241,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/241,chatglm_llm.py文件下第129行先加载了一次chatglm模型,第143行又加载了一次,87 -88,[BUG] similarity_search_with_score_by_vector函数返回多个doc时的score结果错误,2023-05-06.0252,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/252,**问题描述 / Problem Description**,88 -89,可以再建一个交流群吗,这个群满了进不去。,2023-05-06.0255,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/255,上午应该已经在readme里更新过了,如果不能添加可能是网页缓存问题,可以试试看直接扫描img/qr_code_12.jpg,89 -90,请问这是什么错误哇?KeyError: 'serialized_input',2023-05-06.0257,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/257,运行“python webui.py” 后这是什么错误?怎么解决啊?,90 -91,修改哪里的代码,可以再cpu上跑?,2023-05-06.0258,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/258,**问题描述 / Problem Description**,91 -92,ModuleNotFoundError: No module named 'modelscope',2023-05-07.0266,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/266,安装这个,92 -93,加载lora微调模型时,lora参数加载成功,但显示模型未成功加载?,2023-05-08.0270,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/270,什么原因呀?,93 -94,[BUG] 运行webui.py报错:name 'EMBEDDING_DEVICE' is not defined,2023-05-08.0274,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/274,解决了,我修改model_config时候把这个变量改错了,94 -95,基于ptuning训练完成,新老模型都进行了加载,但是只有新的,2023-05-08.0280,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/280,licitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision.,95 -96,[BUG] 使用chatyuan模型时,对话Error,has no attribute 'stream_chat',2023-05-08.0282,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/282,**问题描述 / Problem Description**,96 -97,chaglm调用过程中 _call提示有一个 stop,2023-05-09.0286,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/286,**功能描述 / Feature Description**,97 -98,Logger._log() got an unexpected keyword argument 'end',2023-05-10.0295,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/295,使用cli_demo的时候,加载一个普通txt文件,输入问题后,报错:“TypeError: Logger._log() got an unexpected keyword argument 'end'”,98 -99,[BUG] 请问可以解释下这个FAISS.similarity_search_with_score_by_vector = similarity_search_with_score_by_vector的目的吗,2023-05-10.0296,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/296,我不太明白这个库自己写的similarity_search_with_score_by_vector方法做的事情,因为langchain原版的similarity_search_with_score_by_vector只是search faiss之后把返回的topk句子组合起来。我觉得原版理解起来没什么问题,但是这个库里自己写的我就没太看明白多做了什么其他的事情,因为没有注释。,99 -100,[BUG] Windows下上传中文文件名文件,faiss无法生成向量数据库文件,2023-05-11.0318,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/318,**问题描述 / Problem Description**,100 -101,cli_demo中的流式输出能否接着前一答案输出?,2023-05-11.0320,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/320,现有流式输出结果样式为:,101 -102,内网部署时网页无法加载,能否增加离线静态资源,2023-05-12.0326,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/326,内网部署时网页无法加载,能否增加离线静态资源,102 -103,我想把文件字符的编码格式改为encoding='utf-8'在哪修改呢,因为会有ascii codec can't decode byte报错,2023-05-14.0360,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/360,上传中文的txt文件时报错,编码格式为utf-8,103 -104,Batches的进度条是在哪里设置的?能否关闭显示?,2023-05-15.0366,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/366,"使用cli_demo.py进行命令行测试时,每句回答前都有个Batches的进度条",104 -105,ImportError: dlopen: cannot load any more object with static TLS or Segmentation fault,2023-05-15.0368,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/368,**问题描述 / Problem Description**,105 -106,读取PDF时报错,2023-05-16.0373,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/373,在Colab上执行cli_demo.py时,在路径文件夹里放了pdf文件,在加载的过程中会显示错误,然后无法加载PDF文件,106 -107,[BUG] webui报错 InvalidURL,2023-05-16.0375,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/375,python 版本:3.8.16,107 -108,[FEATURE] 如果让回答不包含出处,应该怎么处理,2023-05-16.0380,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/380,**功能描述 / Feature Description**,108 -109,加载PDF文件时,出现 unsupported colorspace for 'png',2023-05-16.0381,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/381,**问题描述 / Problem Description**,109 -110,'ascii' codec can't encode characters in position 14-44: ordinal not in range(128) 经典bug,2023-05-16.0382,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/382,添加了知识库之后进行对话,之后再新增知识库就会出现这个问题。,110 -111,微信群人数超过200了,扫码进不去了,群主可以再创建一个新群吗,2023-05-17.0391,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/391,**功能描述 / Feature Description**,111 -112,TypeError: 'ListDocsResponse' object is not subscriptable,2023-05-17.0393,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/393,应该是用remain_docs.code和remain_docs.data吧?吗?,112 -113,[BUG] 加载chatglm模型报错:'NoneType' object has no attribute 'message_types_by_name',2023-05-17.0398,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/398,**问题描述 / Problem Description**,113 -114,[BUG] 执行 python webui.py 没有报错,但是ui界面提示 Something went wrong Expecting value: line 1 column 1 (char 0,2023-05-18.0399,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/399,**环境配置**,114 -115,启动后调用api接口正常,过一会就不断的爆出 Since the angle classifier is not initialized,2023-05-18.0404,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/404,**问题描述 / Problem Description**,115 -116,[BUG] write_check_file方法中,open函数未指定编码,2023-05-18.0408,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/408,"def write_check_file(filepath, docs):",116 -117,导入的PDF中存在图片,有大概率出现 “unsupported colorspace for 'png'”异常,2023-05-18.0409,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/409,"pix = fitz.Pixmap(doc, img[0])",117 -118,请问流程图是用什么软件画的,2023-05-18.0410,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/410,draw.io,118 -119,mac 加载模型失败,2023-05-19.0417,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/417,Explicitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision.,119 -120,使用GPU本地运行知识库问答,提问第一个问题出现异常。,2023-05-20.0419,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/419,配置文件model_config.py为:,120 -121,想加入讨论群,2023-05-20.0420,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/420,OK,121 -122,有没有直接调用LLM的API,目前只有知识库的API?,2023-05-22.0426,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/426,-------------------------------------------------------------------------------,122 -123,上传文件后出现 ERROR __init__() got an unexpected keyword argument 'autodetect_encoding',2023-05-22.0428,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/428,"上传文件后出现这个问题:ERROR 2023-05-22 11:46:19,568-1d: __init__() got an unexpected keyword argument 'autodetect_encoding'",123 -124,想问下README中用到的流程图用什么软件画的,2023-05-22.0431,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/431,**功能描述 / Feature Description**,124 -125,No matching distribution found for langchain==0.0.174,2023-05-23.0436,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/436,ERROR: Could not find a version that satisfies the requirement langchain==0.0.174 ,125 -126,[FEATURE] bing是必须的么?,2023-05-23.0437,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/437,从这个[脚步](https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/configs/model_config.py#L129)里面发现需要申请bing api,如果不申请,纯用模型推理不可吗?,126 -127,同一台环境下部署了5.22号更新的langchain-chatglm v0.1.13和之前的版本,回复速度明显变慢,2023-05-23.0442,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/442,新langchain-chatglm v0.1.13版本速度很慢,127 -128,Error reported during startup,2023-05-23.0443,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/443,Traceback (most recent call last):,128 -129,"ValueError: not enough values to unpack (expected 2, got 1)on of the issue",2023-05-24.0449,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/449,"File "".cache\huggingface\modules\transformers_modules\chatglm-6b-int4\modeling_chatglm.py"", line 1280, in chat",129 -130,[BUG] API部署,流式输出的函数,少了个question,2023-05-24.0451,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/451,**问题描述 / Problem Description**,130 -131,项目结构的简洁性保持,2023-05-24.0454,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/454,**功能描述 / Feature Description**,131 -132,项目群扫码进不去了,2023-05-24.0455,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/455,项目群扫码进不去了,是否可以加一下微信拉我进群,谢谢!微信号:daniel-0527,132 -133,请求拉我入群讨论,海硕一枚,专注于LLM等相关技术,2023-05-24.0461,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/461,**功能描述 / Feature Description**,133 -134,[BUG] chatglm-6b模型报错OSError: Error no file named pytorch_model.bin found in directory /chatGLM/model/model-6b,2023-05-26.0474,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/474,**1、简述:**,134 -135,现在本项目交流群二维码扫描不进去了,需要群主通过,2023-05-27.0478,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/478,现在本项目交流群二维码扫描不进去了,需要群主通过,135 -136,RuntimeError: Only Tensors of floating point and complex dtype can require gradients,2023-05-28.0483,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/483,刚更新了最新版本:,136 -137,"RuntimeError: ""LayerNormKernelImpl"" not implemented for 'Half'",2023-05-28.0484,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/484,"已经解决了 params 只用两个参数 {'trust_remote_code': True, 'torch_dtype': torch.float16}",137 -138,[BUG] 文件未成功加载,请重新上传文件,2023-05-31.0504,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/504,webui.py,138 -139,[BUG] bug 17 ,pdf和pdf为啥还不一样呢?为啥有的pdf能识别?有的pdf识别不了呢?,2023-05-31.0506,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/506,bug 17 ,pdf和pdf为啥还不一样呢?为啥有的pdf能识别?有的pdf识别不了呢?,139 -140,[FEATURE] 简洁阐述功能 / Concise description of the feature,2023-05-31.0513,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/513,**功能描述 / Feature Description**,140 -141,[BUG] webui.py 加载chatglm-6b-int4 失败,2023-06-02.0524,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/524,**问题描述 / Problem Description**,141 -142,[BUG] webui.py 加载chatglm-6b模型异常,2023-06-02.0525,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/525,**问题描述 / Problem Description**,142 -143,增加对chatgpt的embedding和api调用的支持,2023-06-02.0531,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/531,能否支持openai的embedding api和对话的api?,143 -144,[FEATURE] 调整模型下载的位置,2023-06-02.0537,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/537,模型默认下载到 $HOME/.cache/huggingface/,当 C 盘空间不足时无法完成模型的下载。configs/model_config.py 中也没有调整模型位置的参数。,144 -145,[BUG] langchain=0.0.174 出错,2023-06-04.0543,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/543,**问题描述 / Problem Description**,145 -146,[BUG] 更新后加载本地模型路径不正确,2023-06-05.0545,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/545,**问题描述 / Problem Description**,146 -147,SystemError: 8bit 模型需要 CUDA 支持,或者改用量化后模型!,2023-06-06.0550,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/550,"docker 部署后,启动docker,过会儿容器会自动退出,logs报错 SystemError: 8bit 模型需要 CUDA 支持,或者改用量化后模型! [NVIDIA Container Toolkit](https://github.com/NVIDIA/nvidia-container-toolkit) 也已经安装了",147 -148,[BUG] 上传知识库超过1M报错,2023-06-06.0556,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/556,**问题描述 / Problem Description**,148 -149,打开跨域访问后仍然报错,不能请求,2023-06-06.0560,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/560,报错信息:,149 -150,dialogue_answering 里面的代码是不是没有用到?,没有看到调用,2023-06-07.0571,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/571,dialogue_answering 是干啥的,150 -151,[BUG] 响应速度极慢,应从哪里入手优化?48C/128G/8卡,2023-06-07.0573,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/573,运行环境:ubuntu20.04,151 -152,纯CPU环境下运行cli_demo时报错,提示找不到nvcuda.dll,2023-06-08.0576,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/576,本地部署环境是纯CPU,之前的版本在纯CPU环境下能正常运行,但上传本地知识库经常出现encode问题。今天重新git项目后,运行时出现如下问题,请问该如何解决。,152 -153,如何加载本地的embedding模型(text2vec-large-chinese模型文件),2023-06-08.0582,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/582,"因为需要离线部署,所以要把模型放到本地,我修改了chains/local_doc_qa.py中的HuggingFaceEmbeddings(),在其中加了一个cache_folder的参数,保证下载的文件在cache_folder中,model_name是text2vec-large-chinese。如cache_folder='/home/xx/model/text2vec-large-chinese', model_name='text2vec-large-chinese',这样仍然需要联网下载报错,请问大佬如何解决该问题?",153 -154,ChatGLM-6B 在另外服务器安装好了,请问如何修改model.cofnig.py 来使用它的接口呢??,2023-06-09.0588,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/588,我本来想在这加一个api base url 但是运行web.py 发现 还是会去连huggingface 下载模型,154 -155,[BUG] raise partially initialized module 'charset_normalizer' has no attribute 'md__mypyc' when call interface `upload_file`,2023-06-10.0591,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/591,**问题描述 / Problem Description**,155 -156,[BUG] raise OSError: [Errno 101] Network is unreachable when call interface upload_file and upload .pdf files,2023-06-10.0592,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/592,**问题描述 / Problem Description**,156 -157,如果直接用vicuna作为基座大模型,需要修改的地方有哪些?,2023-06-12.0596,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/596,vicuna模型有直接转换好的没有?也就是llama转换之后的vicuna。,157 -158,[BUG] 通过cli.py调用api时抛出AttributeError: 'NoneType' object has no attribute 'get'错误,2023-06-12.0598,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/598,通过`python cli.py start api --ip localhost --port 8001` 命令调用api时,抛出:,158 -159,[BUG] 通过cli.py调用api时直接报错`langchain-ChatGLM: error: unrecognized arguments: start cli`,2023-06-12.0601,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/601,通过python cli.py start cli启动cli_demo时,报错:,159 -160,[BUG] error: unrecognized arguments: --model-dir conf/models/,2023-06-12.0602,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/602,关键字参数修改了吗?有没有文档啊?大佬,160 -161,[BUG] 上传文件全部失败,2023-06-12.0603,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/603,ERROR: Exception in ASGI application,161 -162,[BUG] config 使用 chatyuan 无法启动,2023-06-12.0604,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/604,"""chatyuan"": {",162 -163,使用fashchat api之后,后台报错APIError 如图所示,2023-06-12.0606,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/606,我按照https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/docs/fastchat.md,163 -164,[BUG] 启用上下文关联,每次embedding搜索到的内容都会比前一次多一段,2023-06-13.0613,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/613,**问题描述 / Problem Description**,164 -165,local_doc_qa.py中MyFAISS.from_documents() 这个语句看不太懂。MyFAISS类中没有这个方法,其父类FAISS和VectorStore中也只有from_texts方法[BUG] 简洁阐述问题 / Concise description of the issue,2023-06-14.0619,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/619,local_doc_qa.py中MyFAISS.from_documents() 这个语句看不太懂。MyFAISS类中没有这个方法,其父类FAISS和VectorStore中也只有from_texts方法,165 -166,[BUG] TypeError: similarity_search_with_score_by_vector() got an unexpected keyword argument 'filter',2023-06-14.0624,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/624,**问题描述 / Problem Description**,166 -167,please delete this issue,2023-06-15.0633,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/633,"sorry, incorrect submission. Please remove this issue!",167 -168,[BUG] vue前端镜像构建失败,2023-06-15.0635,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/635,**问题描述 / Problem Description**,168 -169,ChatGLM-6B模型能否回答英文问题?,2023-06-15.0640,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/640,大佬,请问一下,如果本地知识文档是英文,ChatGLM-6B模型能否回答英文问题?不能的话,有没有替代的模型推荐,期待你的回复,谢谢,169 -170,[BUG] 简洁阐述问题 / Concise description of the issue,2023-06-16.0644,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/644,**问题描述 / Problem Description**,170 -171,KeyError: 3224,2023-06-16.0645,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/645,```,171 +,title,file,url,detail,id +0,加油~以及一些建议,2023-03-31.0002,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/2,加油,我认为你的方向是对的。,0 +1,当前的运行环境是什么,windows还是Linux,2023-04-01.0003,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/3,当前的运行环境是什么,windows还是Linux,python是什么版本?,1 +2,请问这是在CLM基础上运行吗?,2023-04-01.0004,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/4,请问是不是需要本地安装好clm并正常运行的情况下,再按文中的步骤执行才能运行起来?,2 +3,[复现问题] 构造 prompt 时从知识库中提取的文字乱码,2023-04-01.0005,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/5,hi,我在尝试复现 README 中的效果,也使用了 ChatGLM-6B 的 README 作为输入文本,但发现从知识库中提取的文字是乱码,导致构造的 prompt 不可用。想了解如何解决这个问题。,3 +4,后面能否加入上下文对话功能?,2023-04-02.0006,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/6,目前的get_wiki_agent_answer函数中已经实现了历史消息传递的功能,后面我再确认一下是否有langchain中model调用过程中是否传递了chat_history。,4 +5,请问:纯cpu可以吗?,2023-04-03.0007,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/7,很酷的实现,极大地开拓了我的眼界!很顺利的在gpu机器上运行了,5 +6,运行报错:AttributeError: 'NoneType' object has no attribute 'message_types_by_name',2023-04-03.0008,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/8,报错:,6 +7,运行环境:GPU需要多大的?,2023-04-03.0009,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/9,如果按照THUDM/ChatGLM-6B的说法,使用的GPU大小应该在13GB左右,但运行脚本后,占用了24GB还不够。,7 +8,请问本地知识的格式是什么?,2023-04-03.0010,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/10,已测试格式包括docx、md文件中的文本信息,具体格式可以参考 [langchain文档](https://python.langchain.com/en/latest/modules/indexes/document_loaders/examples/unstructured_file.html?highlight=pdf#),8 +9,24G的显存还是爆掉了,是否支持双卡运行,2023-04-03.0011,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/11,RuntimeError: CUDA out of memory. Tried to allocate 96.00 MiB (GPU 0; 23.70 GiB total capacity; 22.18 GiB already allocated; 12.75 MiB free; 22.18 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF,9 +10,你怎么知道embeddings方式和模型训练时候的方式是一样的?,2023-04-03.0012,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/12,embedding和LLM的方式不用一致,embedding能够解决语义检索的需求就行。这个项目里用到embedding是在对本地知识建立索引和对问句转换成向量的过程。,10 +11,是否能提供本地知识文件的格式?,2023-04-04.0013,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/13,是否能提供本地知识文件的格式?,11 +12,是否可以像清华原版跑在8G一以下的卡?,2023-04-04.0016,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/16,是否可以像清华原版跑在8G一以下的卡?我的8G卡爆显存了🤣🤣🤣,12 +13,请教一下langchain协调使用向量库和chatGLM工作的,2023-04-05.0018,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/18,代码里面这段是创建问答模型的,会接入ChatGLM和本地语料的向量库,langchain回答的时候是怎么个优先顺序?先搜向量库,没有再找chatglm么? 还是什么机制?,13 +14,在mac m2max上抛出了ValueError: 150001 is not in list这个异常,2023-04-05.0019,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/19,我把chatglm_llm.py加载模型的代码改成如下,14 +15,程序运行后一直卡住,2023-04-05.0020,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/20,感谢作者的付出,不过本人在运行时出现了问题,请大家帮助。,15 +16,问一下chat_history的逻辑,2023-04-06.0022,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/22,感谢开源。,16 +17,为什么每次运行都会loading checkpoint,2023-04-06.0023,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/23,我把这个embeding模型下载到本地后,无法正常启动。,17 +18,本地知识文件能否上传一些示例?,2023-04-06.0025,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/25,如题,怎么构造知识文件,效果更好?能否提供一个样例,18 +19,What version of you are using?,2023-04-06.0026,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/26,"Hi Panda, I saw the `pip install -r requirements` command in README, and want to confirm you are using python2 or python3? because my pip and pip3 version are all is 22.3.",19 +20,有兴趣交流本项目应用的朋友可以加一下微信群,2023-04-07.0027,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/27,![IMG_1630](https://user-images.githubusercontent.com/5668498/230533162-8b9bfcdd-249c-4efe-b066-4f9ba2ce9f23.jpeg),20 +21,本地知识越多,回答时检索的时间是否会越长,2023-04-07.0029,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/29,是的 因为需要进行向量匹配检索,21 +22,爲啥最後還是報錯 哭。。,2023-04-07.0030,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/30,Failed to import transformers.models.t5.configuration_t5 because of the following error (look up to see,22 +23,对话到第二次的时候就报错UnicodeDecodeError: 'utf-8' codec can't decode,2023-04-07.0031,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/31,对话第一次是没问题的,模型返回输出后又给到请输入你的问题,我再输入问题就报错,23 +24,用的in4的量化版本,推理的时候显示需要申请10Gb的显存,2023-04-07.0033,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/33,"File ""/root/.cache/huggingface/modules/transformers_modules/chatglm-6b-int4-qe/modeling_chatglm.py"", line 581, in forward",24 +25,使用colab运行,python3.9,提示包导入有问题,2023-04-07.0034,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/34,"from ._util import is_directory, is_path",25 +26,运行失败,Loading checkpoint未达到100%被kill了,请问下是什么原因?,2023-04-07.0035,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/35,日志如下:,26 +27,弄了个交流群,自己弄好多细节不会,大家技术讨论 加connection-image 我来拉你,2023-04-08.0036,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/36,自己搞好多不清楚的,一起来弄吧。。准备搞个部署问题的解决文档出来,27 +28,Error using the new version with langchain,2023-04-09.0043,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/43,Error with the new changes:,28 +29,程序报错torch.cuda.OutOfMemoryError如何解决?,2023-04-10.0044,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/44,报错详细信息如下:,29 +30,qa的训练数据格式是如何设置的,2023-04-10.0045,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/45,本项目不是使用微调的方式,所以并不涉及到训练过程。,30 +31,The FileType.UNK file type is not supported in partition. 解决办法,2023-04-10.0046,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/46,ValueError: Invalid file /home/yawu/Documents/langchain-ChatGLM-master/data. The FileType.UNK file type is not supported in partition.,31 +32,如何读取多个txt文档?,2023-04-10.0047,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/47,如题,请教一下如何读取多个txt文档?示例代码中只给了读一个文档的案例,这个input我换成string之后也只能指定一个文档,无法用通配符指定多个文档,也无法传入多个文件路径的列表。,32 +33,nltk package unable to either download or load local nltk_data folder,2023-04-10.0049,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/49,I'm running this project on an offline Windows Server environment so I download the Punkt and averaged_perceptron_tagger tokenizer in this directory:,33 +34,requirements.txt中需要指定langchain版本,2023-04-11.0055,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/55,langchain版本0.116下无法引入RetrievalQA,需要指定更高版本(0.136版本下无问题),34 +35,Demo演示无法给出输出内容,2023-04-12.0059,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/59,你好,测试了项目自带新闻稿示例和自行上传的一个文本,可以加载进去,但是无法给出答案,请问属于什么情况,如何解决,谢谢。PS: 1、今天早上刚下载全部代码;2、硬件服务器满足要求;3、按操作说明正常操作。,35 +36,群人数过多无法进群,求帮忙拉进群,2023-04-12.0061,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/61,您好,您的群人数超过了200人,目前无法通过二维码加群,请问您方便加我微信拉我进群吗?万分感谢,36 +37,群人数已满,求大佬拉入群,2023-04-12.0062,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/62,已在README中更新拉群二维码,37 +38,requirements中langchain版本错误,2023-04-12.0065,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/65,langchain版本应该是0.0.12而不是0.0.120,38 +39,Linux : Searchd in,2023-04-13.0068,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/68,import nltk,39 +40,No sentence-transformers model found,2023-04-13.0069,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/69,加载不了这个模型,错误原因是找不到这个模型,但是路径是配置好了的,40 +41,Error loading punkt: ",58 +59,为啥放到方法调用会出错,这个怎么处理?,2023-04-20.0150,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/150,```python,59 +60,No sentence-transformers model found with name C:\Users\Administrator/.cache\torch\sentence_transformers\GanymedeNil_text2vec-large-chinese. Creating a new one with MEAN pooling.,2023-04-21.0154,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/154,卡在这块很久是正常现象吗,60 +61,微信群需要邀请才能加入,2023-04-21.0155,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/155,RT,给个个人联系方式白,61 +62,No sentence-transformers model found with name GanymedeNil/text2vec-large-chinese. Creating a new one with MEAN pooling,2023-04-21.0156,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/156,ls GanymedeNil/text2vec-large-chinese,62 +63,embedding会加载两次,2023-04-23.0159,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/159,你好,为什么要这样设置呢,这样会加载两次呀。,63 +64,扫二维码加的那个群,群成员满了进不去了,2023-04-23.0160,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/160,如题,64 +65,执行python3 cli_demo.py 报错AttributeError: 'NoneType' object has no attribute 'chat',2023-04-24.0163,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/163,"刚开始怀疑是内存不足问题,换成int4,int4-qe也不行,有人知道是什么原因吗",65 +66,匹配得分,2023-04-24.0167,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/167,在示例cli_demo.py中返回的匹配文本没有对应的score,可以加上这个feature吗,66 +67,大佬有计划往web_ui.py加入打字机功能吗,2023-04-25.0170,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/170,目前在载入了知识库后,单张V100 32G在回答垂直领域的问题时也需要20S以上,没有打字机逐字输出的使用体验还是比较煎熬的....,67 +68,Is it possible to use a verctorDB for the embedings?,2023-04-25.0171,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/171,"when I play, I have to load the local data again and again when to start. I wonder if it is possible to use",68 +69,请问通过lora训练官方模型得到的微调模型文件该如何加载?,2023-04-25.0173,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/173,通过lora训练的方式得到以下文件:,69 +70,from langchain.chains import RetrievalQA的代码在哪里?,2023-04-25.0174,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/174,local_doc_qa.py,70 +71,哪里有knowledge_based_chatglm.py文件?怎么找不到了??是被替换成cli_demo.py文件了吗?,2023-04-26.0175,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/175,哪里有knowledge_based_chatglm.py文件?怎么找不到了??是被替换成cli_demo.py文件了吗?,71 +72,AttributeError: 'Chatbot' object has no attribute 'value',2023-04-26.0177,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/177,Traceback (most recent call last):,72 +73,控制台调api.py报警告,2023-04-26.0178,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/178,"you must pass the application as an import string to enable ""reload"" or ""workers""",73 +74,如何加入群聊,2023-04-27.0183,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/183,微信群超过200人了,需要邀请,如何加入呢?,74 +75,如何将Chatglm和本地知识相结合,2023-04-27.0185,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/185,您好,我想请教一下怎么才能让知识库匹配到的文本和chatglm生成的相结合,而不是说如果没搜索到,就说根据已知信息无法回答该问题,谢谢,75 +76,一点建议,2023-04-27.0189,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/189,1.weiui的get_vector_store方法里面添加一个判断以兼容gradio版本导致的上传异常,76 +77,windows环境下,按照教程,配置好conda环境,git完项目,修改完模型路径相关内容后,运行demo报错缺少,2023-04-28.0194,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/194,报错代码如下:,77 +78,ValueError: too many values to unpack (expected 2),2023-04-28.0198,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/198,"When i tried to use the non-streaming, `ValueError: too many values to unpack (expected 2)` error came out.",78 +79,加载doc后覆盖原本知识,2023-04-28.0201,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/201,加载较大量级的私有知识库后,原本的知识会被覆盖,79 +80,自定义知识库回答效果很差,2023-04-28.0203,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/203,"请问加了自定义知识库知识库,回答效果很差,是因为数据量太小的原因么",80 +81,python310下,安装pycocotools失败,提示低版本cython,实际已安装高版本,2023-04-29.0208,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/208,RT,纯离线环境安装,依赖安装的十分艰难,最后碰到pycocotools,始终无法安装上,求教方法!,81 +82,[FEATURE] 支持 RWKV 模型(目前已有 pip package & rwkv.cpp 等等),2023-05-01.0216,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/216,您好,我是 RWKV 的作者,介绍见:https://zhuanlan.zhihu.com/p/626083366,82 +83,[BUG] 为啥主机/服务器不联网不能正常启动服务?,2023-05-02.0220,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/220,**问题描述 / Problem Description**,83 +84,[BUG] 简洁阐述问题 / Concise description of the issue,2023-05-03.0222,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/222,**local variable 'torch' referenced before assignment**,84 +85,不支持txt文件的中文输入,2023-05-04.0235,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/235,"vs_path, _ = local_doc_qa.init_knowledge_vector_store(filepath)",85 +86,文件均未成功加载,请检查依赖包或替换为其他文件再次上传。 文件未成功加载,请重新上传文件,2023-05-05.0237,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/237,请大佬帮忙解决,谢谢!,86 +87,[BUG] 使用多卡时chatglm模型加载两次,2023-05-05.0241,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/241,chatglm_llm.py文件下第129行先加载了一次chatglm模型,第143行又加载了一次,87 +88,[BUG] similarity_search_with_score_by_vector函数返回多个doc时的score结果错误,2023-05-06.0252,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/252,**问题描述 / Problem Description**,88 +89,可以再建一个交流群吗,这个群满了进不去。,2023-05-06.0255,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/255,上午应该已经在readme里更新过了,如果不能添加可能是网页缓存问题,可以试试看直接扫描img/qr_code_12.jpg,89 +90,请问这是什么错误哇?KeyError: 'serialized_input',2023-05-06.0257,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/257,运行“python webui.py” 后这是什么错误?怎么解决啊?,90 +91,修改哪里的代码,可以再cpu上跑?,2023-05-06.0258,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/258,**问题描述 / Problem Description**,91 +92,ModuleNotFoundError: No module named 'modelscope',2023-05-07.0266,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/266,安装这个,92 +93,加载lora微调模型时,lora参数加载成功,但显示模型未成功加载?,2023-05-08.0270,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/270,什么原因呀?,93 +94,[BUG] 运行webui.py报错:name 'EMBEDDING_DEVICE' is not defined,2023-05-08.0274,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/274,解决了,我修改model_config时候把这个变量改错了,94 +95,基于ptuning训练完成,新老模型都进行了加载,但是只有新的,2023-05-08.0280,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/280,licitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision.,95 +96,[BUG] 使用chatyuan模型时,对话Error,has no attribute 'stream_chat',2023-05-08.0282,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/282,**问题描述 / Problem Description**,96 +97,chaglm调用过程中 _call提示有一个 stop,2023-05-09.0286,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/286,**功能描述 / Feature Description**,97 +98,Logger._log() got an unexpected keyword argument 'end',2023-05-10.0295,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/295,使用cli_demo的时候,加载一个普通txt文件,输入问题后,报错:“TypeError: Logger._log() got an unexpected keyword argument 'end'”,98 +99,[BUG] 请问可以解释下这个FAISS.similarity_search_with_score_by_vector = similarity_search_with_score_by_vector的目的吗,2023-05-10.0296,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/296,我不太明白这个库自己写的similarity_search_with_score_by_vector方法做的事情,因为langchain原版的similarity_search_with_score_by_vector只是search faiss之后把返回的topk句子组合起来。我觉得原版理解起来没什么问题,但是这个库里自己写的我就没太看明白多做了什么其他的事情,因为没有注释。,99 +100,[BUG] Windows下上传中文文件名文件,faiss无法生成向量数据库文件,2023-05-11.0318,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/318,**问题描述 / Problem Description**,100 +101,cli_demo中的流式输出能否接着前一答案输出?,2023-05-11.0320,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/320,现有流式输出结果样式为:,101 +102,内网部署时网页无法加载,能否增加离线静态资源,2023-05-12.0326,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/326,内网部署时网页无法加载,能否增加离线静态资源,102 +103,我想把文件字符的编码格式改为encoding='utf-8'在哪修改呢,因为会有ascii codec can't decode byte报错,2023-05-14.0360,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/360,上传中文的txt文件时报错,编码格式为utf-8,103 +104,Batches的进度条是在哪里设置的?能否关闭显示?,2023-05-15.0366,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/366,"使用cli_demo.py进行命令行测试时,每句回答前都有个Batches的进度条",104 +105,ImportError: dlopen: cannot load any more object with static TLS or Segmentation fault,2023-05-15.0368,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/368,**问题描述 / Problem Description**,105 +106,读取PDF时报错,2023-05-16.0373,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/373,在Colab上执行cli_demo.py时,在路径文件夹里放了pdf文件,在加载的过程中会显示错误,然后无法加载PDF文件,106 +107,[BUG] webui报错 InvalidURL,2023-05-16.0375,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/375,python 版本:3.8.16,107 +108,[FEATURE] 如果让回答不包含出处,应该怎么处理,2023-05-16.0380,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/380,**功能描述 / Feature Description**,108 +109,加载PDF文件时,出现 unsupported colorspace for 'png',2023-05-16.0381,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/381,**问题描述 / Problem Description**,109 +110,'ascii' codec can't encode characters in position 14-44: ordinal not in range(128) 经典bug,2023-05-16.0382,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/382,添加了知识库之后进行对话,之后再新增知识库就会出现这个问题。,110 +111,微信群人数超过200了,扫码进不去了,群主可以再创建一个新群吗,2023-05-17.0391,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/391,**功能描述 / Feature Description**,111 +112,TypeError: 'ListDocsResponse' object is not subscriptable,2023-05-17.0393,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/393,应该是用remain_docs.code和remain_docs.data吧?吗?,112 +113,[BUG] 加载chatglm模型报错:'NoneType' object has no attribute 'message_types_by_name',2023-05-17.0398,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/398,**问题描述 / Problem Description**,113 +114,[BUG] 执行 python webui.py 没有报错,但是ui界面提示 Something went wrong Expecting value: line 1 column 1 (char 0,2023-05-18.0399,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/399,**环境配置**,114 +115,启动后调用api接口正常,过一会就不断的爆出 Since the angle classifier is not initialized,2023-05-18.0404,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/404,**问题描述 / Problem Description**,115 +116,[BUG] write_check_file方法中,open函数未指定编码,2023-05-18.0408,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/408,"def write_check_file(filepath, docs):",116 +117,导入的PDF中存在图片,有大概率出现 “unsupported colorspace for 'png'”异常,2023-05-18.0409,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/409,"pix = fitz.Pixmap(doc, img[0])",117 +118,请问流程图是用什么软件画的,2023-05-18.0410,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/410,draw.io,118 +119,mac 加载模型失败,2023-05-19.0417,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/417,Explicitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision.,119 +120,使用GPU本地运行知识库问答,提问第一个问题出现异常。,2023-05-20.0419,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/419,配置文件model_config.py为:,120 +121,想加入讨论群,2023-05-20.0420,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/420,OK,121 +122,有没有直接调用LLM的API,目前只有知识库的API?,2023-05-22.0426,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/426,-------------------------------------------------------------------------------,122 +123,上传文件后出现 ERROR __init__() got an unexpected keyword argument 'autodetect_encoding',2023-05-22.0428,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/428,"上传文件后出现这个问题:ERROR 2023-05-22 11:46:19,568-1d: __init__() got an unexpected keyword argument 'autodetect_encoding'",123 +124,想问下README中用到的流程图用什么软件画的,2023-05-22.0431,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/431,**功能描述 / Feature Description**,124 +125,No matching distribution found for langchain==0.0.174,2023-05-23.0436,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/436,ERROR: Could not find a version that satisfies the requirement langchain==0.0.174 ,125 +126,[FEATURE] bing是必须的么?,2023-05-23.0437,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/437,从这个[脚步](https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/configs/model_config.py#L129)里面发现需要申请bing api,如果不申请,纯用模型推理不可吗?,126 +127,同一台环境下部署了5.22号更新的langchain-chatglm v0.1.13和之前的版本,回复速度明显变慢,2023-05-23.0442,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/442,新langchain-chatglm v0.1.13版本速度很慢,127 +128,Error reported during startup,2023-05-23.0443,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/443,Traceback (most recent call last):,128 +129,"ValueError: not enough values to unpack (expected 2, got 1)on of the issue",2023-05-24.0449,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/449,"File "".cache\huggingface\modules\transformers_modules\chatglm-6b-int4\modeling_chatglm.py"", line 1280, in chat",129 +130,[BUG] API部署,流式输出的函数,少了个question,2023-05-24.0451,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/451,**问题描述 / Problem Description**,130 +131,项目结构的简洁性保持,2023-05-24.0454,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/454,**功能描述 / Feature Description**,131 +132,项目群扫码进不去了,2023-05-24.0455,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/455,项目群扫码进不去了,是否可以加一下微信拉我进群,谢谢!微信号:daniel-0527,132 +133,请求拉我入群讨论,海硕一枚,专注于LLM等相关技术,2023-05-24.0461,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/461,**功能描述 / Feature Description**,133 +134,[BUG] chatglm-6b模型报错OSError: Error no file named pytorch_model.bin found in directory /chatGLM/model/model-6b,2023-05-26.0474,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/474,**1、简述:**,134 +135,现在本项目交流群二维码扫描不进去了,需要群主通过,2023-05-27.0478,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/478,现在本项目交流群二维码扫描不进去了,需要群主通过,135 +136,RuntimeError: Only Tensors of floating point and complex dtype can require gradients,2023-05-28.0483,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/483,刚更新了最新版本:,136 +137,"RuntimeError: ""LayerNormKernelImpl"" not implemented for 'Half'",2023-05-28.0484,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/484,"已经解决了 params 只用两个参数 {'trust_remote_code': True, 'torch_dtype': torch.float16}",137 +138,[BUG] 文件未成功加载,请重新上传文件,2023-05-31.0504,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/504,webui.py,138 +139,[BUG] bug 17 ,pdf和pdf为啥还不一样呢?为啥有的pdf能识别?有的pdf识别不了呢?,2023-05-31.0506,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/506,bug 17 ,pdf和pdf为啥还不一样呢?为啥有的pdf能识别?有的pdf识别不了呢?,139 +140,[FEATURE] 简洁阐述功能 / Concise description of the feature,2023-05-31.0513,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/513,**功能描述 / Feature Description**,140 +141,[BUG] webui.py 加载chatglm-6b-int4 失败,2023-06-02.0524,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/524,**问题描述 / Problem Description**,141 +142,[BUG] webui.py 加载chatglm-6b模型异常,2023-06-02.0525,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/525,**问题描述 / Problem Description**,142 +143,增加对chatgpt的embedding和api调用的支持,2023-06-02.0531,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/531,能否支持openai的embedding api和对话的api?,143 +144,[FEATURE] 调整模型下载的位置,2023-06-02.0537,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/537,模型默认下载到 $HOME/.cache/huggingface/,当 C 盘空间不足时无法完成模型的下载。configs/model_config.py 中也没有调整模型位置的参数。,144 +145,[BUG] langchain=0.0.174 出错,2023-06-04.0543,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/543,**问题描述 / Problem Description**,145 +146,[BUG] 更新后加载本地模型路径不正确,2023-06-05.0545,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/545,**问题描述 / Problem Description**,146 +147,SystemError: 8bit 模型需要 CUDA 支持,或者改用量化后模型!,2023-06-06.0550,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/550,"docker 部署后,启动docker,过会儿容器会自动退出,logs报错 SystemError: 8bit 模型需要 CUDA 支持,或者改用量化后模型! [NVIDIA Container Toolkit](https://github.com/NVIDIA/nvidia-container-toolkit) 也已经安装了",147 +148,[BUG] 上传知识库超过1M报错,2023-06-06.0556,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/556,**问题描述 / Problem Description**,148 +149,打开跨域访问后仍然报错,不能请求,2023-06-06.0560,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/560,报错信息:,149 +150,dialogue_answering 里面的代码是不是没有用到?,没有看到调用,2023-06-07.0571,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/571,dialogue_answering 是干啥的,150 +151,[BUG] 响应速度极慢,应从哪里入手优化?48C/128G/8卡,2023-06-07.0573,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/573,运行环境:ubuntu20.04,151 +152,纯CPU环境下运行cli_demo时报错,提示找不到nvcuda.dll,2023-06-08.0576,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/576,本地部署环境是纯CPU,之前的版本在纯CPU环境下能正常运行,但上传本地知识库经常出现encode问题。今天重新git项目后,运行时出现如下问题,请问该如何解决。,152 +153,如何加载本地的embedding模型(text2vec-large-chinese模型文件),2023-06-08.0582,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/582,"因为需要离线部署,所以要把模型放到本地,我修改了chains/local_doc_qa.py中的HuggingFaceEmbeddings(),在其中加了一个cache_folder的参数,保证下载的文件在cache_folder中,model_name是text2vec-large-chinese。如cache_folder='/home/xx/model/text2vec-large-chinese', model_name='text2vec-large-chinese',这样仍然需要联网下载报错,请问大佬如何解决该问题?",153 +154,ChatGLM-6B 在另外服务器安装好了,请问如何修改model.cofnig.py 来使用它的接口呢??,2023-06-09.0588,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/588,我本来想在这加一个api base url 但是运行web.py 发现 还是会去连huggingface 下载模型,154 +155,[BUG] raise partially initialized module 'charset_normalizer' has no attribute 'md__mypyc' when call interface `upload_file`,2023-06-10.0591,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/591,**问题描述 / Problem Description**,155 +156,[BUG] raise OSError: [Errno 101] Network is unreachable when call interface upload_file and upload .pdf files,2023-06-10.0592,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/592,**问题描述 / Problem Description**,156 +157,如果直接用vicuna作为基座大模型,需要修改的地方有哪些?,2023-06-12.0596,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/596,vicuna模型有直接转换好的没有?也就是llama转换之后的vicuna。,157 +158,[BUG] 通过cli.py调用api时抛出AttributeError: 'NoneType' object has no attribute 'get'错误,2023-06-12.0598,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/598,通过`python cli.py start api --ip localhost --port 8001` 命令调用api时,抛出:,158 +159,[BUG] 通过cli.py调用api时直接报错`langchain-ChatGLM: error: unrecognized arguments: start cli`,2023-06-12.0601,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/601,通过python cli.py start cli启动cli_demo时,报错:,159 +160,[BUG] error: unrecognized arguments: --model-dir conf/models/,2023-06-12.0602,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/602,关键字参数修改了吗?有没有文档啊?大佬,160 +161,[BUG] 上传文件全部失败,2023-06-12.0603,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/603,ERROR: Exception in ASGI application,161 +162,[BUG] config 使用 chatyuan 无法启动,2023-06-12.0604,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/604,"""chatyuan"": {",162 +163,使用fashchat api之后,后台报错APIError 如图所示,2023-06-12.0606,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/606,我按照https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/docs/fastchat.md,163 +164,[BUG] 启用上下文关联,每次embedding搜索到的内容都会比前一次多一段,2023-06-13.0613,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/613,**问题描述 / Problem Description**,164 +165,local_doc_qa.py中MyFAISS.from_documents() 这个语句看不太懂。MyFAISS类中没有这个方法,其父类FAISS和VectorStore中也只有from_texts方法[BUG] 简洁阐述问题 / Concise description of the issue,2023-06-14.0619,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/619,local_doc_qa.py中MyFAISS.from_documents() 这个语句看不太懂。MyFAISS类中没有这个方法,其父类FAISS和VectorStore中也只有from_texts方法,165 +166,[BUG] TypeError: similarity_search_with_score_by_vector() got an unexpected keyword argument 'filter',2023-06-14.0624,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/624,**问题描述 / Problem Description**,166 +167,please delete this issue,2023-06-15.0633,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/633,"sorry, incorrect submission. Please remove this issue!",167 +168,[BUG] vue前端镜像构建失败,2023-06-15.0635,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/635,**问题描述 / Problem Description**,168 +169,ChatGLM-6B模型能否回答英文问题?,2023-06-15.0640,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/640,大佬,请问一下,如果本地知识文档是英文,ChatGLM-6B模型能否回答英文问题?不能的话,有没有替代的模型推荐,期待你的回复,谢谢,169 +170,[BUG] 简洁阐述问题 / Concise description of the issue,2023-06-16.0644,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/644,**问题描述 / Problem Description**,170 +171,KeyError: 3224,2023-06-16.0645,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/645,```,171 diff --git a/knowledge_base/samples/isssues_merge/langchain-ChatGLM_closed.jsonl b/knowledge_base/samples/isssues_merge/langchain-ChatGLM_closed.jsonl index 237cd9d2..fd2040e1 100644 --- a/knowledge_base/samples/isssues_merge/langchain-ChatGLM_closed.jsonl +++ b/knowledge_base/samples/isssues_merge/langchain-ChatGLM_closed.jsonl @@ -1,172 +1,172 @@ -{"title": "加油~以及一些建议", "file": "2023-03-31.0002", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/2", "detail": "加油,我认为你的方向是对的。", "id": 0} -{"title": "当前的运行环境是什么,windows还是Linux", "file": "2023-04-01.0003", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/3", "detail": "当前的运行环境是什么,windows还是Linux,python是什么版本?", "id": 1} -{"title": "请问这是在CLM基础上运行吗?", "file": "2023-04-01.0004", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/4", "detail": "请问是不是需要本地安装好clm并正常运行的情况下,再按文中的步骤执行才能运行起来?", "id": 2} -{"title": "[复现问题] 构造 prompt 时从知识库中提取的文字乱码", "file": "2023-04-01.0005", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/5", "detail": "hi,我在尝试复现 README 中的效果,也使用了 ChatGLM-6B 的 README 作为输入文本,但发现从知识库中提取的文字是乱码,导致构造的 prompt 不可用。想了解如何解决这个问题。", "id": 3} -{"title": "后面能否加入上下文对话功能?", "file": "2023-04-02.0006", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/6", "detail": "目前的get_wiki_agent_answer函数中已经实现了历史消息传递的功能,后面我再确认一下是否有langchain中model调用过程中是否传递了chat_history。", "id": 4} -{"title": "请问:纯cpu可以吗?", "file": "2023-04-03.0007", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/7", "detail": "很酷的实现,极大地开拓了我的眼界!很顺利的在gpu机器上运行了", "id": 5} -{"title": "运行报错:AttributeError: 'NoneType' object has no attribute 'message_types_by_name'", "file": "2023-04-03.0008", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/8", "detail": "报错:", "id": 6} -{"title": "运行环境:GPU需要多大的?", "file": "2023-04-03.0009", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/9", "detail": "如果按照THUDM/ChatGLM-6B的说法,使用的GPU大小应该在13GB左右,但运行脚本后,占用了24GB还不够。", "id": 7} -{"title": "请问本地知识的格式是什么?", "file": "2023-04-03.0010", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/10", "detail": "已测试格式包括docx、md文件中的文本信息,具体格式可以参考 [langchain文档](https://python.langchain.com/en/latest/modules/indexes/document_loaders/examples/unstructured_file.html?highlight=pdf#)", "id": 8} -{"title": "24G的显存还是爆掉了,是否支持双卡运行", "file": "2023-04-03.0011", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/11", "detail": "RuntimeError: CUDA out of memory. Tried to allocate 96.00 MiB (GPU 0; 23.70 GiB total capacity; 22.18 GiB already allocated; 12.75 MiB free; 22.18 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF", "id": 9} -{"title": "你怎么知道embeddings方式和模型训练时候的方式是一样的?", "file": "2023-04-03.0012", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/12", "detail": "embedding和LLM的方式不用一致,embedding能够解决语义检索的需求就行。这个项目里用到embedding是在对本地知识建立索引和对问句转换成向量的过程。", "id": 10} -{"title": "是否能提供本地知识文件的格式?", "file": "2023-04-04.0013", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/13", "detail": "是否能提供本地知识文件的格式?", "id": 11} -{"title": "是否可以像清华原版跑在8G一以下的卡?", "file": "2023-04-04.0016", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/16", "detail": "是否可以像清华原版跑在8G一以下的卡?我的8G卡爆显存了🤣🤣🤣", "id": 12} -{"title": "请教一下langchain协调使用向量库和chatGLM工作的", "file": "2023-04-05.0018", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/18", "detail": "代码里面这段是创建问答模型的,会接入ChatGLM和本地语料的向量库,langchain回答的时候是怎么个优先顺序?先搜向量库,没有再找chatglm么? 还是什么机制?", "id": 13} -{"title": "在mac m2max上抛出了ValueError: 150001 is not in list这个异常", "file": "2023-04-05.0019", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/19", "detail": "我把chatglm_llm.py加载模型的代码改成如下", "id": 14} -{"title": "程序运行后一直卡住", "file": "2023-04-05.0020", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/20", "detail": "感谢作者的付出,不过本人在运行时出现了问题,请大家帮助。", "id": 15} -{"title": "问一下chat_history的逻辑", "file": "2023-04-06.0022", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/22", "detail": "感谢开源。", "id": 16} -{"title": "为什么每次运行都会loading checkpoint", "file": "2023-04-06.0023", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/23", "detail": "我把这个embeding模型下载到本地后,无法正常启动。", "id": 17} -{"title": "本地知识文件能否上传一些示例?", "file": "2023-04-06.0025", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/25", "detail": "如题,怎么构造知识文件,效果更好?能否提供一个样例", "id": 18} -{"title": "What version of you are using?", "file": "2023-04-06.0026", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/26", "detail": "Hi Panda, I saw the `pip install -r requirements` command in README, and want to confirm you are using python2 or python3? because my pip and pip3 version are all is 22.3.", "id": 19} -{"title": "有兴趣交流本项目应用的朋友可以加一下微信群", "file": "2023-04-07.0027", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/27", "detail": "![IMG_1630](https://user-images.githubusercontent.com/5668498/230533162-8b9bfcdd-249c-4efe-b066-4f9ba2ce9f23.jpeg)", "id": 20} -{"title": "本地知识越多,回答时检索的时间是否会越长", "file": "2023-04-07.0029", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/29", "detail": "是的 因为需要进行向量匹配检索", "id": 21} -{"title": "爲啥最後還是報錯 哭。。", "file": "2023-04-07.0030", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/30", "detail": "Failed to import transformers.models.t5.configuration_t5 because of the following error (look up to see", "id": 22} -{"title": "对话到第二次的时候就报错UnicodeDecodeError: 'utf-8' codec can't decode", "file": "2023-04-07.0031", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/31", "detail": "对话第一次是没问题的,模型返回输出后又给到请输入你的问题,我再输入问题就报错", "id": 23} -{"title": "用的in4的量化版本,推理的时候显示需要申请10Gb的显存", "file": "2023-04-07.0033", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/33", "detail": "File \"/root/.cache/huggingface/modules/transformers_modules/chatglm-6b-int4-qe/modeling_chatglm.py\", line 581, in forward", "id": 24} -{"title": "使用colab运行,python3.9,提示包导入有问题", "file": "2023-04-07.0034", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/34", "detail": "from ._util import is_directory, is_path", "id": 25} -{"title": "运行失败,Loading checkpoint未达到100%被kill了,请问下是什么原因?", "file": "2023-04-07.0035", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/35", "detail": "日志如下:", "id": 26} -{"title": "弄了个交流群,自己弄好多细节不会,大家技术讨论 加connection-image 我来拉你", "file": "2023-04-08.0036", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/36", "detail": "自己搞好多不清楚的,一起来弄吧。。准备搞个部署问题的解决文档出来", "id": 27} -{"title": "Error using the new version with langchain", "file": "2023-04-09.0043", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/43", "detail": "Error with the new changes:", "id": 28} -{"title": "程序报错torch.cuda.OutOfMemoryError如何解决?", "file": "2023-04-10.0044", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/44", "detail": "报错详细信息如下:", "id": 29} -{"title": "qa的训练数据格式是如何设置的", "file": "2023-04-10.0045", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/45", "detail": "本项目不是使用微调的方式,所以并不涉及到训练过程。", "id": 30} -{"title": "The FileType.UNK file type is not supported in partition. 解决办法", "file": "2023-04-10.0046", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/46", "detail": "ValueError: Invalid file /home/yawu/Documents/langchain-ChatGLM-master/data. The FileType.UNK file type is not supported in partition.", "id": 31} -{"title": "如何读取多个txt文档?", "file": "2023-04-10.0047", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/47", "detail": "如题,请教一下如何读取多个txt文档?示例代码中只给了读一个文档的案例,这个input我换成string之后也只能指定一个文档,无法用通配符指定多个文档,也无法传入多个文件路径的列表。", "id": 32} -{"title": "nltk package unable to either download or load local nltk_data folder", "file": "2023-04-10.0049", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/49", "detail": "I'm running this project on an offline Windows Server environment so I download the Punkt and averaged_perceptron_tagger tokenizer in this directory:", "id": 33} -{"title": "requirements.txt中需要指定langchain版本", "file": "2023-04-11.0055", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/55", "detail": "langchain版本0.116下无法引入RetrievalQA,需要指定更高版本(0.136版本下无问题)", "id": 34} -{"title": "Demo演示无法给出输出内容", "file": "2023-04-12.0059", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/59", "detail": "你好,测试了项目自带新闻稿示例和自行上传的一个文本,可以加载进去,但是无法给出答案,请问属于什么情况,如何解决,谢谢。PS: 1、今天早上刚下载全部代码;2、硬件服务器满足要求;3、按操作说明正常操作。", "id": 35} -{"title": "群人数过多无法进群,求帮忙拉进群", "file": "2023-04-12.0061", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/61", "detail": "您好,您的群人数超过了200人,目前无法通过二维码加群,请问您方便加我微信拉我进群吗?万分感谢", "id": 36} -{"title": "群人数已满,求大佬拉入群", "file": "2023-04-12.0062", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/62", "detail": "已在README中更新拉群二维码", "id": 37} -{"title": "requirements中langchain版本错误", "file": "2023-04-12.0065", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/65", "detail": "langchain版本应该是0.0.12而不是0.0.120", "id": 38} -{"title": "Linux : Searchd in", "file": "2023-04-13.0068", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/68", "detail": "import nltk", "id": 39} -{"title": "No sentence-transformers model found", "file": "2023-04-13.0069", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/69", "detail": "加载不了这个模型,错误原因是找不到这个模型,但是路径是配置好了的", "id": 40} -{"title": "Error loading punkt: ", "id": 58} -{"title": "为啥放到方法调用会出错,这个怎么处理?", "file": "2023-04-20.0150", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/150", "detail": "```python", "id": 59} -{"title": "No sentence-transformers model found with name C:\\Users\\Administrator/.cache\\torch\\sentence_transformers\\GanymedeNil_text2vec-large-chinese. Creating a new one with MEAN pooling.", "file": "2023-04-21.0154", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/154", "detail": "卡在这块很久是正常现象吗", "id": 60} -{"title": "微信群需要邀请才能加入", "file": "2023-04-21.0155", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/155", "detail": "RT,给个个人联系方式白", "id": 61} -{"title": "No sentence-transformers model found with name GanymedeNil/text2vec-large-chinese. Creating a new one with MEAN pooling", "file": "2023-04-21.0156", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/156", "detail": "ls GanymedeNil/text2vec-large-chinese", "id": 62} -{"title": "embedding会加载两次", "file": "2023-04-23.0159", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/159", "detail": "你好,为什么要这样设置呢,这样会加载两次呀。", "id": 63} -{"title": "扫二维码加的那个群,群成员满了进不去了", "file": "2023-04-23.0160", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/160", "detail": "如题", "id": 64} -{"title": "执行python3 cli_demo.py 报错AttributeError: 'NoneType' object has no attribute 'chat'", "file": "2023-04-24.0163", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/163", "detail": "刚开始怀疑是内存不足问题,换成int4,int4-qe也不行,有人知道是什么原因吗", "id": 65} -{"title": "匹配得分", "file": "2023-04-24.0167", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/167", "detail": "在示例cli_demo.py中返回的匹配文本没有对应的score,可以加上这个feature吗", "id": 66} -{"title": "大佬有计划往web_ui.py加入打字机功能吗", "file": "2023-04-25.0170", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/170", "detail": "目前在载入了知识库后,单张V100 32G在回答垂直领域的问题时也需要20S以上,没有打字机逐字输出的使用体验还是比较煎熬的....", "id": 67} -{"title": "Is it possible to use a verctorDB for the embedings?", "file": "2023-04-25.0171", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/171", "detail": "when I play, I have to load the local data again and again when to start. I wonder if it is possible to use", "id": 68} -{"title": "请问通过lora训练官方模型得到的微调模型文件该如何加载?", "file": "2023-04-25.0173", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/173", "detail": "通过lora训练的方式得到以下文件:", "id": 69} -{"title": "from langchain.chains import RetrievalQA的代码在哪里?", "file": "2023-04-25.0174", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/174", "detail": "local_doc_qa.py", "id": 70} -{"title": "哪里有knowledge_based_chatglm.py文件?怎么找不到了??是被替换成cli_demo.py文件了吗?", "file": "2023-04-26.0175", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/175", "detail": "哪里有knowledge_based_chatglm.py文件?怎么找不到了??是被替换成cli_demo.py文件了吗?", "id": 71} -{"title": "AttributeError: 'Chatbot' object has no attribute 'value'", "file": "2023-04-26.0177", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/177", "detail": "Traceback (most recent call last):", "id": 72} -{"title": "控制台调api.py报警告", "file": "2023-04-26.0178", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/178", "detail": "you must pass the application as an import string to enable \"reload\" or \"workers\"", "id": 73} -{"title": "如何加入群聊", "file": "2023-04-27.0183", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/183", "detail": "微信群超过200人了,需要邀请,如何加入呢?", "id": 74} -{"title": "如何将Chatglm和本地知识相结合", "file": "2023-04-27.0185", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/185", "detail": "您好,我想请教一下怎么才能让知识库匹配到的文本和chatglm生成的相结合,而不是说如果没搜索到,就说根据已知信息无法回答该问题,谢谢", "id": 75} -{"title": "一点建议", "file": "2023-04-27.0189", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/189", "detail": "1.weiui的get_vector_store方法里面添加一个判断以兼容gradio版本导致的上传异常", "id": 76} -{"title": "windows环境下,按照教程,配置好conda环境,git完项目,修改完模型路径相关内容后,运行demo报错缺少", "file": "2023-04-28.0194", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/194", "detail": "报错代码如下:", "id": 77} -{"title": "ValueError: too many values to unpack (expected 2)", "file": "2023-04-28.0198", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/198", "detail": "When i tried to use the non-streaming, `ValueError: too many values to unpack (expected 2)` error came out.", "id": 78} -{"title": "加载doc后覆盖原本知识", "file": "2023-04-28.0201", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/201", "detail": "加载较大量级的私有知识库后,原本的知识会被覆盖", "id": 79} -{"title": "自定义知识库回答效果很差", "file": "2023-04-28.0203", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/203", "detail": "请问加了自定义知识库知识库,回答效果很差,是因为数据量太小的原因么", "id": 80} -{"title": "python310下,安装pycocotools失败,提示低版本cython,实际已安装高版本", "file": "2023-04-29.0208", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/208", "detail": "RT,纯离线环境安装,依赖安装的十分艰难,最后碰到pycocotools,始终无法安装上,求教方法!", "id": 81} -{"title": "[FEATURE] 支持 RWKV 模型(目前已有 pip package & rwkv.cpp 等等)", "file": "2023-05-01.0216", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/216", "detail": "您好,我是 RWKV 的作者,介绍见:https://zhuanlan.zhihu.com/p/626083366", "id": 82} -{"title": "[BUG] 为啥主机/服务器不联网不能正常启动服务?", "file": "2023-05-02.0220", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/220", "detail": "**问题描述 / Problem Description**", "id": 83} -{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-05-03.0222", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/222", "detail": "**local variable 'torch' referenced before assignment**", "id": 84} -{"title": "不支持txt文件的中文输入", "file": "2023-05-04.0235", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/235", "detail": "vs_path, _ = local_doc_qa.init_knowledge_vector_store(filepath)", "id": 85} -{"title": "文件均未成功加载,请检查依赖包或替换为其他文件再次上传。 文件未成功加载,请重新上传文件", "file": "2023-05-05.0237", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/237", "detail": "请大佬帮忙解决,谢谢!", "id": 86} -{"title": "[BUG] 使用多卡时chatglm模型加载两次", "file": "2023-05-05.0241", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/241", "detail": "chatglm_llm.py文件下第129行先加载了一次chatglm模型,第143行又加载了一次", "id": 87} -{"title": "[BUG] similarity_search_with_score_by_vector函数返回多个doc时的score结果错误", "file": "2023-05-06.0252", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/252", "detail": "**问题描述 / Problem Description**", "id": 88} -{"title": "可以再建一个交流群吗,这个群满了进不去。", "file": "2023-05-06.0255", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/255", "detail": "上午应该已经在readme里更新过了,如果不能添加可能是网页缓存问题,可以试试看直接扫描img/qr_code_12.jpg", "id": 89} -{"title": "请问这是什么错误哇?KeyError: 'serialized_input'", "file": "2023-05-06.0257", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/257", "detail": "运行“python webui.py” 后这是什么错误?怎么解决啊?", "id": 90} -{"title": "修改哪里的代码,可以再cpu上跑?", "file": "2023-05-06.0258", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/258", "detail": "**问题描述 / Problem Description**", "id": 91} -{"title": "ModuleNotFoundError: No module named 'modelscope'", "file": "2023-05-07.0266", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/266", "detail": "安装这个", "id": 92} -{"title": "加载lora微调模型时,lora参数加载成功,但显示模型未成功加载?", "file": "2023-05-08.0270", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/270", "detail": "什么原因呀?", "id": 93} -{"title": "[BUG] 运行webui.py报错:name 'EMBEDDING_DEVICE' is not defined", "file": "2023-05-08.0274", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/274", "detail": "解决了,我修改model_config时候把这个变量改错了", "id": 94} -{"title": "基于ptuning训练完成,新老模型都进行了加载,但是只有新的", "file": "2023-05-08.0280", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/280", "detail": "licitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision.", "id": 95} -{"title": "[BUG] 使用chatyuan模型时,对话Error,has no attribute 'stream_chat'", "file": "2023-05-08.0282", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/282", "detail": "**问题描述 / Problem Description**", "id": 96} -{"title": "chaglm调用过程中 _call提示有一个 stop", "file": "2023-05-09.0286", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/286", "detail": "**功能描述 / Feature Description**", "id": 97} -{"title": "Logger._log() got an unexpected keyword argument 'end'", "file": "2023-05-10.0295", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/295", "detail": "使用cli_demo的时候,加载一个普通txt文件,输入问题后,报错:“TypeError: Logger._log() got an unexpected keyword argument 'end'”", "id": 98} -{"title": "[BUG] 请问可以解释下这个FAISS.similarity_search_with_score_by_vector = similarity_search_with_score_by_vector的目的吗", "file": "2023-05-10.0296", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/296", "detail": "我不太明白这个库自己写的similarity_search_with_score_by_vector方法做的事情,因为langchain原版的similarity_search_with_score_by_vector只是search faiss之后把返回的topk句子组合起来。我觉得原版理解起来没什么问题,但是这个库里自己写的我就没太看明白多做了什么其他的事情,因为没有注释。", "id": 99} -{"title": "[BUG] Windows下上传中文文件名文件,faiss无法生成向量数据库文件", "file": "2023-05-11.0318", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/318", "detail": "**问题描述 / Problem Description**", "id": 100} -{"title": "cli_demo中的流式输出能否接着前一答案输出?", "file": "2023-05-11.0320", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/320", "detail": "现有流式输出结果样式为:", "id": 101} -{"title": "内网部署时网页无法加载,能否增加离线静态资源", "file": "2023-05-12.0326", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/326", "detail": "内网部署时网页无法加载,能否增加离线静态资源", "id": 102} -{"title": "我想把文件字符的编码格式改为encoding='utf-8'在哪修改呢,因为会有ascii codec can't decode byte报错", "file": "2023-05-14.0360", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/360", "detail": "上传中文的txt文件时报错,编码格式为utf-8", "id": 103} -{"title": "Batches的进度条是在哪里设置的?能否关闭显示?", "file": "2023-05-15.0366", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/366", "detail": "使用cli_demo.py进行命令行测试时,每句回答前都有个Batches的进度条", "id": 104} -{"title": "ImportError: dlopen: cannot load any more object with static TLS or Segmentation fault", "file": "2023-05-15.0368", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/368", "detail": "**问题描述 / Problem Description**", "id": 105} -{"title": "读取PDF时报错", "file": "2023-05-16.0373", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/373", "detail": "在Colab上执行cli_demo.py时,在路径文件夹里放了pdf文件,在加载的过程中会显示错误,然后无法加载PDF文件", "id": 106} -{"title": "[BUG] webui报错 InvalidURL", "file": "2023-05-16.0375", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/375", "detail": "python 版本:3.8.16", "id": 107} -{"title": "[FEATURE] 如果让回答不包含出处,应该怎么处理", "file": "2023-05-16.0380", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/380", "detail": "**功能描述 / Feature Description**", "id": 108} -{"title": "加载PDF文件时,出现 unsupported colorspace for 'png'", "file": "2023-05-16.0381", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/381", "detail": "**问题描述 / Problem Description**", "id": 109} -{"title": "'ascii' codec can't encode characters in position 14-44: ordinal not in range(128) 经典bug", "file": "2023-05-16.0382", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/382", "detail": "添加了知识库之后进行对话,之后再新增知识库就会出现这个问题。", "id": 110} -{"title": "微信群人数超过200了,扫码进不去了,群主可以再创建一个新群吗", "file": "2023-05-17.0391", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/391", "detail": "**功能描述 / Feature Description**", "id": 111} -{"title": "TypeError: 'ListDocsResponse' object is not subscriptable", "file": "2023-05-17.0393", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/393", "detail": "应该是用remain_docs.code和remain_docs.data吧?吗?", "id": 112} -{"title": "[BUG] 加载chatglm模型报错:'NoneType' object has no attribute 'message_types_by_name'", "file": "2023-05-17.0398", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/398", "detail": "**问题描述 / Problem Description**", "id": 113} -{"title": "[BUG] 执行 python webui.py 没有报错,但是ui界面提示 Something went wrong Expecting value: line 1 column 1 (char 0", "file": "2023-05-18.0399", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/399", "detail": "**环境配置**", "id": 114} -{"title": "启动后调用api接口正常,过一会就不断的爆出 Since the angle classifier is not initialized", "file": "2023-05-18.0404", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/404", "detail": "**问题描述 / Problem Description**", "id": 115} -{"title": "[BUG] write_check_file方法中,open函数未指定编码", "file": "2023-05-18.0408", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/408", "detail": "def write_check_file(filepath, docs):", "id": 116} -{"title": "导入的PDF中存在图片,有大概率出现 “unsupported colorspace for 'png'”异常", "file": "2023-05-18.0409", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/409", "detail": "pix = fitz.Pixmap(doc, img[0])", "id": 117} -{"title": "请问流程图是用什么软件画的", "file": "2023-05-18.0410", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/410", "detail": "draw.io", "id": 118} -{"title": "mac 加载模型失败", "file": "2023-05-19.0417", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/417", "detail": "Explicitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision.", "id": 119} -{"title": "使用GPU本地运行知识库问答,提问第一个问题出现异常。", "file": "2023-05-20.0419", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/419", "detail": "配置文件model_config.py为:", "id": 120} -{"title": "想加入讨论群", "file": "2023-05-20.0420", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/420", "detail": "OK", "id": 121} -{"title": "有没有直接调用LLM的API,目前只有知识库的API?", "file": "2023-05-22.0426", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/426", "detail": "-------------------------------------------------------------------------------", "id": 122} -{"title": "上传文件后出现 ERROR __init__() got an unexpected keyword argument 'autodetect_encoding'", "file": "2023-05-22.0428", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/428", "detail": "上传文件后出现这个问题:ERROR 2023-05-22 11:46:19,568-1d: __init__() got an unexpected keyword argument 'autodetect_encoding'", "id": 123} -{"title": "想问下README中用到的流程图用什么软件画的", "file": "2023-05-22.0431", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/431", "detail": "**功能描述 / Feature Description**", "id": 124} -{"title": "No matching distribution found for langchain==0.0.174", "file": "2023-05-23.0436", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/436", "detail": "ERROR: Could not find a version that satisfies the requirement langchain==0.0.174 ", "id": 125} -{"title": "[FEATURE] bing是必须的么?", "file": "2023-05-23.0437", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/437", "detail": "从这个[脚步](https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/configs/model_config.py#L129)里面发现需要申请bing api,如果不申请,纯用模型推理不可吗?", "id": 126} -{"title": "同一台环境下部署了5.22号更新的langchain-chatglm v0.1.13和之前的版本,回复速度明显变慢", "file": "2023-05-23.0442", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/442", "detail": "新langchain-chatglm v0.1.13版本速度很慢", "id": 127} -{"title": "Error reported during startup", "file": "2023-05-23.0443", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/443", "detail": "Traceback (most recent call last):", "id": 128} -{"title": "ValueError: not enough values to unpack (expected 2, got 1)on of the issue", "file": "2023-05-24.0449", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/449", "detail": "File \".cache\\huggingface\\modules\\transformers_modules\\chatglm-6b-int4\\modeling_chatglm.py\", line 1280, in chat", "id": 129} -{"title": "[BUG] API部署,流式输出的函数,少了个question", "file": "2023-05-24.0451", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/451", "detail": "**问题描述 / Problem Description**", "id": 130} -{"title": "项目结构的简洁性保持", "file": "2023-05-24.0454", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/454", "detail": "**功能描述 / Feature Description**", "id": 131} -{"title": "项目群扫码进不去了", "file": "2023-05-24.0455", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/455", "detail": "项目群扫码进不去了,是否可以加一下微信拉我进群,谢谢!微信号:daniel-0527", "id": 132} -{"title": "请求拉我入群讨论,海硕一枚,专注于LLM等相关技术", "file": "2023-05-24.0461", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/461", "detail": "**功能描述 / Feature Description**", "id": 133} -{"title": "[BUG] chatglm-6b模型报错OSError: Error no file named pytorch_model.bin found in directory /chatGLM/model/model-6b", "file": "2023-05-26.0474", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/474", "detail": "**1、简述:**", "id": 134} -{"title": "现在本项目交流群二维码扫描不进去了,需要群主通过", "file": "2023-05-27.0478", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/478", "detail": "现在本项目交流群二维码扫描不进去了,需要群主通过", "id": 135} -{"title": "RuntimeError: Only Tensors of floating point and complex dtype can require gradients", "file": "2023-05-28.0483", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/483", "detail": "刚更新了最新版本:", "id": 136} -{"title": "RuntimeError: \"LayerNormKernelImpl\" not implemented for 'Half'", "file": "2023-05-28.0484", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/484", "detail": "已经解决了 params 只用两个参数 {'trust_remote_code': True, 'torch_dtype': torch.float16}", "id": 137} -{"title": "[BUG] 文件未成功加载,请重新上传文件", "file": "2023-05-31.0504", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/504", "detail": "webui.py", "id": 138} -{"title": "[BUG] bug 17 ,pdf和pdf为啥还不一样呢?为啥有的pdf能识别?有的pdf识别不了呢?", "file": "2023-05-31.0506", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/506", "detail": "bug 17 ,pdf和pdf为啥还不一样呢?为啥有的pdf能识别?有的pdf识别不了呢?", "id": 139} -{"title": "[FEATURE] 简洁阐述功能 / Concise description of the feature", "file": "2023-05-31.0513", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/513", "detail": "**功能描述 / Feature Description**", "id": 140} -{"title": "[BUG] webui.py 加载chatglm-6b-int4 失败", "file": "2023-06-02.0524", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/524", "detail": "**问题描述 / Problem Description**", "id": 141} -{"title": "[BUG] webui.py 加载chatglm-6b模型异常", "file": "2023-06-02.0525", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/525", "detail": "**问题描述 / Problem Description**", "id": 142} -{"title": "增加对chatgpt的embedding和api调用的支持", "file": "2023-06-02.0531", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/531", "detail": "能否支持openai的embedding api和对话的api?", "id": 143} -{"title": "[FEATURE] 调整模型下载的位置", "file": "2023-06-02.0537", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/537", "detail": "模型默认下载到 $HOME/.cache/huggingface/,当 C 盘空间不足时无法完成模型的下载。configs/model_config.py 中也没有调整模型位置的参数。", "id": 144} -{"title": "[BUG] langchain=0.0.174 出错", "file": "2023-06-04.0543", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/543", "detail": "**问题描述 / Problem Description**", "id": 145} -{"title": "[BUG] 更新后加载本地模型路径不正确", "file": "2023-06-05.0545", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/545", "detail": "**问题描述 / Problem Description**", "id": 146} -{"title": "SystemError: 8bit 模型需要 CUDA 支持,或者改用量化后模型!", "file": "2023-06-06.0550", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/550", "detail": "docker 部署后,启动docker,过会儿容器会自动退出,logs报错 SystemError: 8bit 模型需要 CUDA 支持,或者改用量化后模型! [NVIDIA Container Toolkit](https://github.com/NVIDIA/nvidia-container-toolkit) 也已经安装了", "id": 147} -{"title": "[BUG] 上传知识库超过1M报错", "file": "2023-06-06.0556", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/556", "detail": "**问题描述 / Problem Description**", "id": 148} -{"title": "打开跨域访问后仍然报错,不能请求", "file": "2023-06-06.0560", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/560", "detail": "报错信息:", "id": 149} -{"title": "dialogue_answering 里面的代码是不是没有用到?,没有看到调用", "file": "2023-06-07.0571", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/571", "detail": "dialogue_answering 是干啥的", "id": 150} -{"title": "[BUG] 响应速度极慢,应从哪里入手优化?48C/128G/8卡", "file": "2023-06-07.0573", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/573", "detail": "运行环境:ubuntu20.04", "id": 151} -{"title": "纯CPU环境下运行cli_demo时报错,提示找不到nvcuda.dll", "file": "2023-06-08.0576", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/576", "detail": "本地部署环境是纯CPU,之前的版本在纯CPU环境下能正常运行,但上传本地知识库经常出现encode问题。今天重新git项目后,运行时出现如下问题,请问该如何解决。", "id": 152} -{"title": "如何加载本地的embedding模型(text2vec-large-chinese模型文件)", "file": "2023-06-08.0582", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/582", "detail": "因为需要离线部署,所以要把模型放到本地,我修改了chains/local_doc_qa.py中的HuggingFaceEmbeddings(),在其中加了一个cache_folder的参数,保证下载的文件在cache_folder中,model_name是text2vec-large-chinese。如cache_folder='/home/xx/model/text2vec-large-chinese', model_name='text2vec-large-chinese',这样仍然需要联网下载报错,请问大佬如何解决该问题?", "id": 153} -{"title": "ChatGLM-6B 在另外服务器安装好了,请问如何修改model.cofnig.py 来使用它的接口呢??", "file": "2023-06-09.0588", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/588", "detail": "我本来想在这加一个api base url 但是运行web.py 发现 还是会去连huggingface 下载模型", "id": 154} -{"title": "[BUG] raise partially initialized module 'charset_normalizer' has no attribute 'md__mypyc' when call interface `upload_file`", "file": "2023-06-10.0591", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/591", "detail": "**问题描述 / Problem Description**", "id": 155} -{"title": "[BUG] raise OSError: [Errno 101] Network is unreachable when call interface upload_file and upload .pdf files", "file": "2023-06-10.0592", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/592", "detail": "**问题描述 / Problem Description**", "id": 156} -{"title": "如果直接用vicuna作为基座大模型,需要修改的地方有哪些?", "file": "2023-06-12.0596", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/596", "detail": "vicuna模型有直接转换好的没有?也就是llama转换之后的vicuna。", "id": 157} -{"title": "[BUG] 通过cli.py调用api时抛出AttributeError: 'NoneType' object has no attribute 'get'错误", "file": "2023-06-12.0598", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/598", "detail": "通过`python cli.py start api --ip localhost --port 8001` 命令调用api时,抛出:", "id": 158} -{"title": "[BUG] 通过cli.py调用api时直接报错`langchain-ChatGLM: error: unrecognized arguments: start cli`", "file": "2023-06-12.0601", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/601", "detail": "通过python cli.py start cli启动cli_demo时,报错:", "id": 159} -{"title": "[BUG] error: unrecognized arguments: --model-dir conf/models/", "file": "2023-06-12.0602", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/602", "detail": "关键字参数修改了吗?有没有文档啊?大佬", "id": 160} -{"title": "[BUG] 上传文件全部失败", "file": "2023-06-12.0603", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/603", "detail": "ERROR: Exception in ASGI application", "id": 161} -{"title": "[BUG] config 使用 chatyuan 无法启动", "file": "2023-06-12.0604", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/604", "detail": "\"chatyuan\": {", "id": 162} -{"title": "使用fashchat api之后,后台报错APIError 如图所示", "file": "2023-06-12.0606", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/606", "detail": "我按照https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/docs/fastchat.md", "id": 163} -{"title": "[BUG] 启用上下文关联,每次embedding搜索到的内容都会比前一次多一段", "file": "2023-06-13.0613", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/613", "detail": "**问题描述 / Problem Description**", "id": 164} -{"title": "local_doc_qa.py中MyFAISS.from_documents() 这个语句看不太懂。MyFAISS类中没有这个方法,其父类FAISS和VectorStore中也只有from_texts方法[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-06-14.0619", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/619", "detail": "local_doc_qa.py中MyFAISS.from_documents() 这个语句看不太懂。MyFAISS类中没有这个方法,其父类FAISS和VectorStore中也只有from_texts方法", "id": 165} -{"title": "[BUG] TypeError: similarity_search_with_score_by_vector() got an unexpected keyword argument 'filter'", "file": "2023-06-14.0624", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/624", "detail": "**问题描述 / Problem Description**", "id": 166} -{"title": "please delete this issue", "file": "2023-06-15.0633", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/633", "detail": "sorry, incorrect submission. Please remove this issue!", "id": 167} -{"title": "[BUG] vue前端镜像构建失败", "file": "2023-06-15.0635", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/635", "detail": "**问题描述 / Problem Description**", "id": 168} -{"title": "ChatGLM-6B模型能否回答英文问题?", "file": "2023-06-15.0640", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/640", "detail": "大佬,请问一下,如果本地知识文档是英文,ChatGLM-6B模型能否回答英文问题?不能的话,有没有替代的模型推荐,期待你的回复,谢谢", "id": 169} -{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-06-16.0644", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/644", "detail": "**问题描述 / Problem Description**", "id": 170} -{"title": "KeyError: 3224", "file": "2023-06-16.0645", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/645", "detail": "```", "id": 171} +{"title": "加油~以及一些建议", "file": "2023-03-31.0002", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/2", "detail": "加油,我认为你的方向是对的。", "id": 0} +{"title": "当前的运行环境是什么,windows还是Linux", "file": "2023-04-01.0003", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/3", "detail": "当前的运行环境是什么,windows还是Linux,python是什么版本?", "id": 1} +{"title": "请问这是在CLM基础上运行吗?", "file": "2023-04-01.0004", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/4", "detail": "请问是不是需要本地安装好clm并正常运行的情况下,再按文中的步骤执行才能运行起来?", "id": 2} +{"title": "[复现问题] 构造 prompt 时从知识库中提取的文字乱码", "file": "2023-04-01.0005", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/5", "detail": "hi,我在尝试复现 README 中的效果,也使用了 ChatGLM-6B 的 README 作为输入文本,但发现从知识库中提取的文字是乱码,导致构造的 prompt 不可用。想了解如何解决这个问题。", "id": 3} +{"title": "后面能否加入上下文对话功能?", "file": "2023-04-02.0006", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/6", "detail": "目前的get_wiki_agent_answer函数中已经实现了历史消息传递的功能,后面我再确认一下是否有langchain中model调用过程中是否传递了chat_history。", "id": 4} +{"title": "请问:纯cpu可以吗?", "file": "2023-04-03.0007", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/7", "detail": "很酷的实现,极大地开拓了我的眼界!很顺利的在gpu机器上运行了", "id": 5} +{"title": "运行报错:AttributeError: 'NoneType' object has no attribute 'message_types_by_name'", "file": "2023-04-03.0008", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/8", "detail": "报错:", "id": 6} +{"title": "运行环境:GPU需要多大的?", "file": "2023-04-03.0009", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/9", "detail": "如果按照THUDM/ChatGLM-6B的说法,使用的GPU大小应该在13GB左右,但运行脚本后,占用了24GB还不够。", "id": 7} +{"title": "请问本地知识的格式是什么?", "file": "2023-04-03.0010", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/10", "detail": "已测试格式包括docx、md文件中的文本信息,具体格式可以参考 [langchain文档](https://python.langchain.com/en/latest/modules/indexes/document_loaders/examples/unstructured_file.html?highlight=pdf#)", "id": 8} +{"title": "24G的显存还是爆掉了,是否支持双卡运行", "file": "2023-04-03.0011", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/11", "detail": "RuntimeError: CUDA out of memory. Tried to allocate 96.00 MiB (GPU 0; 23.70 GiB total capacity; 22.18 GiB already allocated; 12.75 MiB free; 22.18 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF", "id": 9} +{"title": "你怎么知道embeddings方式和模型训练时候的方式是一样的?", "file": "2023-04-03.0012", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/12", "detail": "embedding和LLM的方式不用一致,embedding能够解决语义检索的需求就行。这个项目里用到embedding是在对本地知识建立索引和对问句转换成向量的过程。", "id": 10} +{"title": "是否能提供本地知识文件的格式?", "file": "2023-04-04.0013", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/13", "detail": "是否能提供本地知识文件的格式?", "id": 11} +{"title": "是否可以像清华原版跑在8G一以下的卡?", "file": "2023-04-04.0016", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/16", "detail": "是否可以像清华原版跑在8G一以下的卡?我的8G卡爆显存了🤣🤣🤣", "id": 12} +{"title": "请教一下langchain协调使用向量库和chatGLM工作的", "file": "2023-04-05.0018", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/18", "detail": "代码里面这段是创建问答模型的,会接入ChatGLM和本地语料的向量库,langchain回答的时候是怎么个优先顺序?先搜向量库,没有再找chatglm么? 还是什么机制?", "id": 13} +{"title": "在mac m2max上抛出了ValueError: 150001 is not in list这个异常", "file": "2023-04-05.0019", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/19", "detail": "我把chatglm_llm.py加载模型的代码改成如下", "id": 14} +{"title": "程序运行后一直卡住", "file": "2023-04-05.0020", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/20", "detail": "感谢作者的付出,不过本人在运行时出现了问题,请大家帮助。", "id": 15} +{"title": "问一下chat_history的逻辑", "file": "2023-04-06.0022", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/22", "detail": "感谢开源。", "id": 16} +{"title": "为什么每次运行都会loading checkpoint", "file": "2023-04-06.0023", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/23", "detail": "我把这个embeding模型下载到本地后,无法正常启动。", "id": 17} +{"title": "本地知识文件能否上传一些示例?", "file": "2023-04-06.0025", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/25", "detail": "如题,怎么构造知识文件,效果更好?能否提供一个样例", "id": 18} +{"title": "What version of you are using?", "file": "2023-04-06.0026", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/26", "detail": "Hi Panda, I saw the `pip install -r requirements` command in README, and want to confirm you are using python2 or python3? because my pip and pip3 version are all is 22.3.", "id": 19} +{"title": "有兴趣交流本项目应用的朋友可以加一下微信群", "file": "2023-04-07.0027", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/27", "detail": "![IMG_1630](https://user-images.githubusercontent.com/5668498/230533162-8b9bfcdd-249c-4efe-b066-4f9ba2ce9f23.jpeg)", "id": 20} +{"title": "本地知识越多,回答时检索的时间是否会越长", "file": "2023-04-07.0029", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/29", "detail": "是的 因为需要进行向量匹配检索", "id": 21} +{"title": "爲啥最後還是報錯 哭。。", "file": "2023-04-07.0030", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/30", "detail": "Failed to import transformers.models.t5.configuration_t5 because of the following error (look up to see", "id": 22} +{"title": "对话到第二次的时候就报错UnicodeDecodeError: 'utf-8' codec can't decode", "file": "2023-04-07.0031", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/31", "detail": "对话第一次是没问题的,模型返回输出后又给到请输入你的问题,我再输入问题就报错", "id": 23} +{"title": "用的in4的量化版本,推理的时候显示需要申请10Gb的显存", "file": "2023-04-07.0033", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/33", "detail": "File \"/root/.cache/huggingface/modules/transformers_modules/chatglm-6b-int4-qe/modeling_chatglm.py\", line 581, in forward", "id": 24} +{"title": "使用colab运行,python3.9,提示包导入有问题", "file": "2023-04-07.0034", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/34", "detail": "from ._util import is_directory, is_path", "id": 25} +{"title": "运行失败,Loading checkpoint未达到100%被kill了,请问下是什么原因?", "file": "2023-04-07.0035", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/35", "detail": "日志如下:", "id": 26} +{"title": "弄了个交流群,自己弄好多细节不会,大家技术讨论 加connection-image 我来拉你", "file": "2023-04-08.0036", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/36", "detail": "自己搞好多不清楚的,一起来弄吧。。准备搞个部署问题的解决文档出来", "id": 27} +{"title": "Error using the new version with langchain", "file": "2023-04-09.0043", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/43", "detail": "Error with the new changes:", "id": 28} +{"title": "程序报错torch.cuda.OutOfMemoryError如何解决?", "file": "2023-04-10.0044", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/44", "detail": "报错详细信息如下:", "id": 29} +{"title": "qa的训练数据格式是如何设置的", "file": "2023-04-10.0045", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/45", "detail": "本项目不是使用微调的方式,所以并不涉及到训练过程。", "id": 30} +{"title": "The FileType.UNK file type is not supported in partition. 解决办法", "file": "2023-04-10.0046", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/46", "detail": "ValueError: Invalid file /home/yawu/Documents/langchain-ChatGLM-master/data. The FileType.UNK file type is not supported in partition.", "id": 31} +{"title": "如何读取多个txt文档?", "file": "2023-04-10.0047", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/47", "detail": "如题,请教一下如何读取多个txt文档?示例代码中只给了读一个文档的案例,这个input我换成string之后也只能指定一个文档,无法用通配符指定多个文档,也无法传入多个文件路径的列表。", "id": 32} +{"title": "nltk package unable to either download or load local nltk_data folder", "file": "2023-04-10.0049", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/49", "detail": "I'm running this project on an offline Windows Server environment so I download the Punkt and averaged_perceptron_tagger tokenizer in this directory:", "id": 33} +{"title": "requirements.txt中需要指定langchain版本", "file": "2023-04-11.0055", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/55", "detail": "langchain版本0.116下无法引入RetrievalQA,需要指定更高版本(0.136版本下无问题)", "id": 34} +{"title": "Demo演示无法给出输出内容", "file": "2023-04-12.0059", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/59", "detail": "你好,测试了项目自带新闻稿示例和自行上传的一个文本,可以加载进去,但是无法给出答案,请问属于什么情况,如何解决,谢谢。PS: 1、今天早上刚下载全部代码;2、硬件服务器满足要求;3、按操作说明正常操作。", "id": 35} +{"title": "群人数过多无法进群,求帮忙拉进群", "file": "2023-04-12.0061", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/61", "detail": "您好,您的群人数超过了200人,目前无法通过二维码加群,请问您方便加我微信拉我进群吗?万分感谢", "id": 36} +{"title": "群人数已满,求大佬拉入群", "file": "2023-04-12.0062", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/62", "detail": "已在README中更新拉群二维码", "id": 37} +{"title": "requirements中langchain版本错误", "file": "2023-04-12.0065", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/65", "detail": "langchain版本应该是0.0.12而不是0.0.120", "id": 38} +{"title": "Linux : Searchd in", "file": "2023-04-13.0068", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/68", "detail": "import nltk", "id": 39} +{"title": "No sentence-transformers model found", "file": "2023-04-13.0069", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/69", "detail": "加载不了这个模型,错误原因是找不到这个模型,但是路径是配置好了的", "id": 40} +{"title": "Error loading punkt: ", "id": 58} +{"title": "为啥放到方法调用会出错,这个怎么处理?", "file": "2023-04-20.0150", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/150", "detail": "```python", "id": 59} +{"title": "No sentence-transformers model found with name C:\\Users\\Administrator/.cache\\torch\\sentence_transformers\\GanymedeNil_text2vec-large-chinese. Creating a new one with MEAN pooling.", "file": "2023-04-21.0154", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/154", "detail": "卡在这块很久是正常现象吗", "id": 60} +{"title": "微信群需要邀请才能加入", "file": "2023-04-21.0155", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/155", "detail": "RT,给个个人联系方式白", "id": 61} +{"title": "No sentence-transformers model found with name GanymedeNil/text2vec-large-chinese. Creating a new one with MEAN pooling", "file": "2023-04-21.0156", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/156", "detail": "ls GanymedeNil/text2vec-large-chinese", "id": 62} +{"title": "embedding会加载两次", "file": "2023-04-23.0159", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/159", "detail": "你好,为什么要这样设置呢,这样会加载两次呀。", "id": 63} +{"title": "扫二维码加的那个群,群成员满了进不去了", "file": "2023-04-23.0160", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/160", "detail": "如题", "id": 64} +{"title": "执行python3 cli_demo.py 报错AttributeError: 'NoneType' object has no attribute 'chat'", "file": "2023-04-24.0163", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/163", "detail": "刚开始怀疑是内存不足问题,换成int4,int4-qe也不行,有人知道是什么原因吗", "id": 65} +{"title": "匹配得分", "file": "2023-04-24.0167", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/167", "detail": "在示例cli_demo.py中返回的匹配文本没有对应的score,可以加上这个feature吗", "id": 66} +{"title": "大佬有计划往web_ui.py加入打字机功能吗", "file": "2023-04-25.0170", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/170", "detail": "目前在载入了知识库后,单张V100 32G在回答垂直领域的问题时也需要20S以上,没有打字机逐字输出的使用体验还是比较煎熬的....", "id": 67} +{"title": "Is it possible to use a verctorDB for the embedings?", "file": "2023-04-25.0171", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/171", "detail": "when I play, I have to load the local data again and again when to start. I wonder if it is possible to use", "id": 68} +{"title": "请问通过lora训练官方模型得到的微调模型文件该如何加载?", "file": "2023-04-25.0173", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/173", "detail": "通过lora训练的方式得到以下文件:", "id": 69} +{"title": "from langchain.chains import RetrievalQA的代码在哪里?", "file": "2023-04-25.0174", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/174", "detail": "local_doc_qa.py", "id": 70} +{"title": "哪里有knowledge_based_chatglm.py文件?怎么找不到了??是被替换成cli_demo.py文件了吗?", "file": "2023-04-26.0175", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/175", "detail": "哪里有knowledge_based_chatglm.py文件?怎么找不到了??是被替换成cli_demo.py文件了吗?", "id": 71} +{"title": "AttributeError: 'Chatbot' object has no attribute 'value'", "file": "2023-04-26.0177", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/177", "detail": "Traceback (most recent call last):", "id": 72} +{"title": "控制台调api.py报警告", "file": "2023-04-26.0178", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/178", "detail": "you must pass the application as an import string to enable \"reload\" or \"workers\"", "id": 73} +{"title": "如何加入群聊", "file": "2023-04-27.0183", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/183", "detail": "微信群超过200人了,需要邀请,如何加入呢?", "id": 74} +{"title": "如何将Chatglm和本地知识相结合", "file": "2023-04-27.0185", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/185", "detail": "您好,我想请教一下怎么才能让知识库匹配到的文本和chatglm生成的相结合,而不是说如果没搜索到,就说根据已知信息无法回答该问题,谢谢", "id": 75} +{"title": "一点建议", "file": "2023-04-27.0189", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/189", "detail": "1.weiui的get_vector_store方法里面添加一个判断以兼容gradio版本导致的上传异常", "id": 76} +{"title": "windows环境下,按照教程,配置好conda环境,git完项目,修改完模型路径相关内容后,运行demo报错缺少", "file": "2023-04-28.0194", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/194", "detail": "报错代码如下:", "id": 77} +{"title": "ValueError: too many values to unpack (expected 2)", "file": "2023-04-28.0198", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/198", "detail": "When i tried to use the non-streaming, `ValueError: too many values to unpack (expected 2)` error came out.", "id": 78} +{"title": "加载doc后覆盖原本知识", "file": "2023-04-28.0201", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/201", "detail": "加载较大量级的私有知识库后,原本的知识会被覆盖", "id": 79} +{"title": "自定义知识库回答效果很差", "file": "2023-04-28.0203", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/203", "detail": "请问加了自定义知识库知识库,回答效果很差,是因为数据量太小的原因么", "id": 80} +{"title": "python310下,安装pycocotools失败,提示低版本cython,实际已安装高版本", "file": "2023-04-29.0208", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/208", "detail": "RT,纯离线环境安装,依赖安装的十分艰难,最后碰到pycocotools,始终无法安装上,求教方法!", "id": 81} +{"title": "[FEATURE] 支持 RWKV 模型(目前已有 pip package & rwkv.cpp 等等)", "file": "2023-05-01.0216", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/216", "detail": "您好,我是 RWKV 的作者,介绍见:https://zhuanlan.zhihu.com/p/626083366", "id": 82} +{"title": "[BUG] 为啥主机/服务器不联网不能正常启动服务?", "file": "2023-05-02.0220", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/220", "detail": "**问题描述 / Problem Description**", "id": 83} +{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-05-03.0222", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/222", "detail": "**local variable 'torch' referenced before assignment**", "id": 84} +{"title": "不支持txt文件的中文输入", "file": "2023-05-04.0235", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/235", "detail": "vs_path, _ = local_doc_qa.init_knowledge_vector_store(filepath)", "id": 85} +{"title": "文件均未成功加载,请检查依赖包或替换为其他文件再次上传。 文件未成功加载,请重新上传文件", "file": "2023-05-05.0237", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/237", "detail": "请大佬帮忙解决,谢谢!", "id": 86} +{"title": "[BUG] 使用多卡时chatglm模型加载两次", "file": "2023-05-05.0241", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/241", "detail": "chatglm_llm.py文件下第129行先加载了一次chatglm模型,第143行又加载了一次", "id": 87} +{"title": "[BUG] similarity_search_with_score_by_vector函数返回多个doc时的score结果错误", "file": "2023-05-06.0252", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/252", "detail": "**问题描述 / Problem Description**", "id": 88} +{"title": "可以再建一个交流群吗,这个群满了进不去。", "file": "2023-05-06.0255", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/255", "detail": "上午应该已经在readme里更新过了,如果不能添加可能是网页缓存问题,可以试试看直接扫描img/qr_code_12.jpg", "id": 89} +{"title": "请问这是什么错误哇?KeyError: 'serialized_input'", "file": "2023-05-06.0257", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/257", "detail": "运行“python webui.py” 后这是什么错误?怎么解决啊?", "id": 90} +{"title": "修改哪里的代码,可以再cpu上跑?", "file": "2023-05-06.0258", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/258", "detail": "**问题描述 / Problem Description**", "id": 91} +{"title": "ModuleNotFoundError: No module named 'modelscope'", "file": "2023-05-07.0266", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/266", "detail": "安装这个", "id": 92} +{"title": "加载lora微调模型时,lora参数加载成功,但显示模型未成功加载?", "file": "2023-05-08.0270", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/270", "detail": "什么原因呀?", "id": 93} +{"title": "[BUG] 运行webui.py报错:name 'EMBEDDING_DEVICE' is not defined", "file": "2023-05-08.0274", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/274", "detail": "解决了,我修改model_config时候把这个变量改错了", "id": 94} +{"title": "基于ptuning训练完成,新老模型都进行了加载,但是只有新的", "file": "2023-05-08.0280", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/280", "detail": "licitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision.", "id": 95} +{"title": "[BUG] 使用chatyuan模型时,对话Error,has no attribute 'stream_chat'", "file": "2023-05-08.0282", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/282", "detail": "**问题描述 / Problem Description**", "id": 96} +{"title": "chaglm调用过程中 _call提示有一个 stop", "file": "2023-05-09.0286", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/286", "detail": "**功能描述 / Feature Description**", "id": 97} +{"title": "Logger._log() got an unexpected keyword argument 'end'", "file": "2023-05-10.0295", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/295", "detail": "使用cli_demo的时候,加载一个普通txt文件,输入问题后,报错:“TypeError: Logger._log() got an unexpected keyword argument 'end'”", "id": 98} +{"title": "[BUG] 请问可以解释下这个FAISS.similarity_search_with_score_by_vector = similarity_search_with_score_by_vector的目的吗", "file": "2023-05-10.0296", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/296", "detail": "我不太明白这个库自己写的similarity_search_with_score_by_vector方法做的事情,因为langchain原版的similarity_search_with_score_by_vector只是search faiss之后把返回的topk句子组合起来。我觉得原版理解起来没什么问题,但是这个库里自己写的我就没太看明白多做了什么其他的事情,因为没有注释。", "id": 99} +{"title": "[BUG] Windows下上传中文文件名文件,faiss无法生成向量数据库文件", "file": "2023-05-11.0318", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/318", "detail": "**问题描述 / Problem Description**", "id": 100} +{"title": "cli_demo中的流式输出能否接着前一答案输出?", "file": "2023-05-11.0320", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/320", "detail": "现有流式输出结果样式为:", "id": 101} +{"title": "内网部署时网页无法加载,能否增加离线静态资源", "file": "2023-05-12.0326", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/326", "detail": "内网部署时网页无法加载,能否增加离线静态资源", "id": 102} +{"title": "我想把文件字符的编码格式改为encoding='utf-8'在哪修改呢,因为会有ascii codec can't decode byte报错", "file": "2023-05-14.0360", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/360", "detail": "上传中文的txt文件时报错,编码格式为utf-8", "id": 103} +{"title": "Batches的进度条是在哪里设置的?能否关闭显示?", "file": "2023-05-15.0366", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/366", "detail": "使用cli_demo.py进行命令行测试时,每句回答前都有个Batches的进度条", "id": 104} +{"title": "ImportError: dlopen: cannot load any more object with static TLS or Segmentation fault", "file": "2023-05-15.0368", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/368", "detail": "**问题描述 / Problem Description**", "id": 105} +{"title": "读取PDF时报错", "file": "2023-05-16.0373", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/373", "detail": "在Colab上执行cli_demo.py时,在路径文件夹里放了pdf文件,在加载的过程中会显示错误,然后无法加载PDF文件", "id": 106} +{"title": "[BUG] webui报错 InvalidURL", "file": "2023-05-16.0375", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/375", "detail": "python 版本:3.8.16", "id": 107} +{"title": "[FEATURE] 如果让回答不包含出处,应该怎么处理", "file": "2023-05-16.0380", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/380", "detail": "**功能描述 / Feature Description**", "id": 108} +{"title": "加载PDF文件时,出现 unsupported colorspace for 'png'", "file": "2023-05-16.0381", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/381", "detail": "**问题描述 / Problem Description**", "id": 109} +{"title": "'ascii' codec can't encode characters in position 14-44: ordinal not in range(128) 经典bug", "file": "2023-05-16.0382", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/382", "detail": "添加了知识库之后进行对话,之后再新增知识库就会出现这个问题。", "id": 110} +{"title": "微信群人数超过200了,扫码进不去了,群主可以再创建一个新群吗", "file": "2023-05-17.0391", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/391", "detail": "**功能描述 / Feature Description**", "id": 111} +{"title": "TypeError: 'ListDocsResponse' object is not subscriptable", "file": "2023-05-17.0393", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/393", "detail": "应该是用remain_docs.code和remain_docs.data吧?吗?", "id": 112} +{"title": "[BUG] 加载chatglm模型报错:'NoneType' object has no attribute 'message_types_by_name'", "file": "2023-05-17.0398", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/398", "detail": "**问题描述 / Problem Description**", "id": 113} +{"title": "[BUG] 执行 python webui.py 没有报错,但是ui界面提示 Something went wrong Expecting value: line 1 column 1 (char 0", "file": "2023-05-18.0399", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/399", "detail": "**环境配置**", "id": 114} +{"title": "启动后调用api接口正常,过一会就不断的爆出 Since the angle classifier is not initialized", "file": "2023-05-18.0404", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/404", "detail": "**问题描述 / Problem Description**", "id": 115} +{"title": "[BUG] write_check_file方法中,open函数未指定编码", "file": "2023-05-18.0408", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/408", "detail": "def write_check_file(filepath, docs):", "id": 116} +{"title": "导入的PDF中存在图片,有大概率出现 “unsupported colorspace for 'png'”异常", "file": "2023-05-18.0409", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/409", "detail": "pix = fitz.Pixmap(doc, img[0])", "id": 117} +{"title": "请问流程图是用什么软件画的", "file": "2023-05-18.0410", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/410", "detail": "draw.io", "id": 118} +{"title": "mac 加载模型失败", "file": "2023-05-19.0417", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/417", "detail": "Explicitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code has been contributed in a newer revision.", "id": 119} +{"title": "使用GPU本地运行知识库问答,提问第一个问题出现异常。", "file": "2023-05-20.0419", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/419", "detail": "配置文件model_config.py为:", "id": 120} +{"title": "想加入讨论群", "file": "2023-05-20.0420", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/420", "detail": "OK", "id": 121} +{"title": "有没有直接调用LLM的API,目前只有知识库的API?", "file": "2023-05-22.0426", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/426", "detail": "-------------------------------------------------------------------------------", "id": 122} +{"title": "上传文件后出现 ERROR __init__() got an unexpected keyword argument 'autodetect_encoding'", "file": "2023-05-22.0428", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/428", "detail": "上传文件后出现这个问题:ERROR 2023-05-22 11:46:19,568-1d: __init__() got an unexpected keyword argument 'autodetect_encoding'", "id": 123} +{"title": "想问下README中用到的流程图用什么软件画的", "file": "2023-05-22.0431", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/431", "detail": "**功能描述 / Feature Description**", "id": 124} +{"title": "No matching distribution found for langchain==0.0.174", "file": "2023-05-23.0436", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/436", "detail": "ERROR: Could not find a version that satisfies the requirement langchain==0.0.174 ", "id": 125} +{"title": "[FEATURE] bing是必须的么?", "file": "2023-05-23.0437", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/437", "detail": "从这个[脚步](https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/configs/model_config.py#L129)里面发现需要申请bing api,如果不申请,纯用模型推理不可吗?", "id": 126} +{"title": "同一台环境下部署了5.22号更新的langchain-chatglm v0.1.13和之前的版本,回复速度明显变慢", "file": "2023-05-23.0442", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/442", "detail": "新langchain-chatglm v0.1.13版本速度很慢", "id": 127} +{"title": "Error reported during startup", "file": "2023-05-23.0443", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/443", "detail": "Traceback (most recent call last):", "id": 128} +{"title": "ValueError: not enough values to unpack (expected 2, got 1)on of the issue", "file": "2023-05-24.0449", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/449", "detail": "File \".cache\\huggingface\\modules\\transformers_modules\\chatglm-6b-int4\\modeling_chatglm.py\", line 1280, in chat", "id": 129} +{"title": "[BUG] API部署,流式输出的函数,少了个question", "file": "2023-05-24.0451", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/451", "detail": "**问题描述 / Problem Description**", "id": 130} +{"title": "项目结构的简洁性保持", "file": "2023-05-24.0454", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/454", "detail": "**功能描述 / Feature Description**", "id": 131} +{"title": "项目群扫码进不去了", "file": "2023-05-24.0455", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/455", "detail": "项目群扫码进不去了,是否可以加一下微信拉我进群,谢谢!微信号:daniel-0527", "id": 132} +{"title": "请求拉我入群讨论,海硕一枚,专注于LLM等相关技术", "file": "2023-05-24.0461", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/461", "detail": "**功能描述 / Feature Description**", "id": 133} +{"title": "[BUG] chatglm-6b模型报错OSError: Error no file named pytorch_model.bin found in directory /chatGLM/model/model-6b", "file": "2023-05-26.0474", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/474", "detail": "**1、简述:**", "id": 134} +{"title": "现在本项目交流群二维码扫描不进去了,需要群主通过", "file": "2023-05-27.0478", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/478", "detail": "现在本项目交流群二维码扫描不进去了,需要群主通过", "id": 135} +{"title": "RuntimeError: Only Tensors of floating point and complex dtype can require gradients", "file": "2023-05-28.0483", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/483", "detail": "刚更新了最新版本:", "id": 136} +{"title": "RuntimeError: \"LayerNormKernelImpl\" not implemented for 'Half'", "file": "2023-05-28.0484", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/484", "detail": "已经解决了 params 只用两个参数 {'trust_remote_code': True, 'torch_dtype': torch.float16}", "id": 137} +{"title": "[BUG] 文件未成功加载,请重新上传文件", "file": "2023-05-31.0504", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/504", "detail": "webui.py", "id": 138} +{"title": "[BUG] bug 17 ,pdf和pdf为啥还不一样呢?为啥有的pdf能识别?有的pdf识别不了呢?", "file": "2023-05-31.0506", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/506", "detail": "bug 17 ,pdf和pdf为啥还不一样呢?为啥有的pdf能识别?有的pdf识别不了呢?", "id": 139} +{"title": "[FEATURE] 简洁阐述功能 / Concise description of the feature", "file": "2023-05-31.0513", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/513", "detail": "**功能描述 / Feature Description**", "id": 140} +{"title": "[BUG] webui.py 加载chatglm-6b-int4 失败", "file": "2023-06-02.0524", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/524", "detail": "**问题描述 / Problem Description**", "id": 141} +{"title": "[BUG] webui.py 加载chatglm-6b模型异常", "file": "2023-06-02.0525", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/525", "detail": "**问题描述 / Problem Description**", "id": 142} +{"title": "增加对chatgpt的embedding和api调用的支持", "file": "2023-06-02.0531", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/531", "detail": "能否支持openai的embedding api和对话的api?", "id": 143} +{"title": "[FEATURE] 调整模型下载的位置", "file": "2023-06-02.0537", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/537", "detail": "模型默认下载到 $HOME/.cache/huggingface/,当 C 盘空间不足时无法完成模型的下载。configs/model_config.py 中也没有调整模型位置的参数。", "id": 144} +{"title": "[BUG] langchain=0.0.174 出错", "file": "2023-06-04.0543", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/543", "detail": "**问题描述 / Problem Description**", "id": 145} +{"title": "[BUG] 更新后加载本地模型路径不正确", "file": "2023-06-05.0545", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/545", "detail": "**问题描述 / Problem Description**", "id": 146} +{"title": "SystemError: 8bit 模型需要 CUDA 支持,或者改用量化后模型!", "file": "2023-06-06.0550", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/550", "detail": "docker 部署后,启动docker,过会儿容器会自动退出,logs报错 SystemError: 8bit 模型需要 CUDA 支持,或者改用量化后模型! [NVIDIA Container Toolkit](https://github.com/NVIDIA/nvidia-container-toolkit) 也已经安装了", "id": 147} +{"title": "[BUG] 上传知识库超过1M报错", "file": "2023-06-06.0556", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/556", "detail": "**问题描述 / Problem Description**", "id": 148} +{"title": "打开跨域访问后仍然报错,不能请求", "file": "2023-06-06.0560", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/560", "detail": "报错信息:", "id": 149} +{"title": "dialogue_answering 里面的代码是不是没有用到?,没有看到调用", "file": "2023-06-07.0571", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/571", "detail": "dialogue_answering 是干啥的", "id": 150} +{"title": "[BUG] 响应速度极慢,应从哪里入手优化?48C/128G/8卡", "file": "2023-06-07.0573", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/573", "detail": "运行环境:ubuntu20.04", "id": 151} +{"title": "纯CPU环境下运行cli_demo时报错,提示找不到nvcuda.dll", "file": "2023-06-08.0576", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/576", "detail": "本地部署环境是纯CPU,之前的版本在纯CPU环境下能正常运行,但上传本地知识库经常出现encode问题。今天重新git项目后,运行时出现如下问题,请问该如何解决。", "id": 152} +{"title": "如何加载本地的embedding模型(text2vec-large-chinese模型文件)", "file": "2023-06-08.0582", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/582", "detail": "因为需要离线部署,所以要把模型放到本地,我修改了chains/local_doc_qa.py中的HuggingFaceEmbeddings(),在其中加了一个cache_folder的参数,保证下载的文件在cache_folder中,model_name是text2vec-large-chinese。如cache_folder='/home/xx/model/text2vec-large-chinese', model_name='text2vec-large-chinese',这样仍然需要联网下载报错,请问大佬如何解决该问题?", "id": 153} +{"title": "ChatGLM-6B 在另外服务器安装好了,请问如何修改model.cofnig.py 来使用它的接口呢??", "file": "2023-06-09.0588", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/588", "detail": "我本来想在这加一个api base url 但是运行web.py 发现 还是会去连huggingface 下载模型", "id": 154} +{"title": "[BUG] raise partially initialized module 'charset_normalizer' has no attribute 'md__mypyc' when call interface `upload_file`", "file": "2023-06-10.0591", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/591", "detail": "**问题描述 / Problem Description**", "id": 155} +{"title": "[BUG] raise OSError: [Errno 101] Network is unreachable when call interface upload_file and upload .pdf files", "file": "2023-06-10.0592", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/592", "detail": "**问题描述 / Problem Description**", "id": 156} +{"title": "如果直接用vicuna作为基座大模型,需要修改的地方有哪些?", "file": "2023-06-12.0596", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/596", "detail": "vicuna模型有直接转换好的没有?也就是llama转换之后的vicuna。", "id": 157} +{"title": "[BUG] 通过cli.py调用api时抛出AttributeError: 'NoneType' object has no attribute 'get'错误", "file": "2023-06-12.0598", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/598", "detail": "通过`python cli.py start api --ip localhost --port 8001` 命令调用api时,抛出:", "id": 158} +{"title": "[BUG] 通过cli.py调用api时直接报错`langchain-ChatGLM: error: unrecognized arguments: start cli`", "file": "2023-06-12.0601", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/601", "detail": "通过python cli.py start cli启动cli_demo时,报错:", "id": 159} +{"title": "[BUG] error: unrecognized arguments: --model-dir conf/models/", "file": "2023-06-12.0602", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/602", "detail": "关键字参数修改了吗?有没有文档啊?大佬", "id": 160} +{"title": "[BUG] 上传文件全部失败", "file": "2023-06-12.0603", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/603", "detail": "ERROR: Exception in ASGI application", "id": 161} +{"title": "[BUG] config 使用 chatyuan 无法启动", "file": "2023-06-12.0604", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/604", "detail": "\"chatyuan\": {", "id": 162} +{"title": "使用fashchat api之后,后台报错APIError 如图所示", "file": "2023-06-12.0606", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/606", "detail": "我按照https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/docs/fastchat.md", "id": 163} +{"title": "[BUG] 启用上下文关联,每次embedding搜索到的内容都会比前一次多一段", "file": "2023-06-13.0613", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/613", "detail": "**问题描述 / Problem Description**", "id": 164} +{"title": "local_doc_qa.py中MyFAISS.from_documents() 这个语句看不太懂。MyFAISS类中没有这个方法,其父类FAISS和VectorStore中也只有from_texts方法[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-06-14.0619", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/619", "detail": "local_doc_qa.py中MyFAISS.from_documents() 这个语句看不太懂。MyFAISS类中没有这个方法,其父类FAISS和VectorStore中也只有from_texts方法", "id": 165} +{"title": "[BUG] TypeError: similarity_search_with_score_by_vector() got an unexpected keyword argument 'filter'", "file": "2023-06-14.0624", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/624", "detail": "**问题描述 / Problem Description**", "id": 166} +{"title": "please delete this issue", "file": "2023-06-15.0633", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/633", "detail": "sorry, incorrect submission. Please remove this issue!", "id": 167} +{"title": "[BUG] vue前端镜像构建失败", "file": "2023-06-15.0635", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/635", "detail": "**问题描述 / Problem Description**", "id": 168} +{"title": "ChatGLM-6B模型能否回答英文问题?", "file": "2023-06-15.0640", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/640", "detail": "大佬,请问一下,如果本地知识文档是英文,ChatGLM-6B模型能否回答英文问题?不能的话,有没有替代的模型推荐,期待你的回复,谢谢", "id": 169} +{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-06-16.0644", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/644", "detail": "**问题描述 / Problem Description**", "id": 170} +{"title": "KeyError: 3224", "file": "2023-06-16.0645", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/645", "detail": "```", "id": 171} diff --git a/knowledge_base/samples/isssues_merge/langchain-ChatGLM_open.csv b/knowledge_base/samples/isssues_merge/langchain-ChatGLM_open.csv index 56ac2ae9..56ba7ca3 100644 --- a/knowledge_base/samples/isssues_merge/langchain-ChatGLM_open.csv +++ b/knowledge_base/samples/isssues_merge/langchain-ChatGLM_open.csv @@ -1,324 +1,324 @@ -,title,file,url,detail,id -0,效果如何优化,2023-04-04.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/14,如图所示,将该项目的README.md和该项目结合后,回答效果并不理想,请问可以从哪些方面进行优化,0 -1,怎么让模型严格根据检索的数据进行回答,减少胡说八道的回答呢,2023-04-04.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/15,举个例子:,1 -2,"When I try to run the `python knowledge_based_chatglm.py`, I got this error in macOS(M1 Max, OS 13.2)",2023-04-07.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/32,```python,2 -3,萌新求教大佬怎么改成AMD显卡或者CPU?,2023-04-10.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/48,把.cuda()去掉就行,3 -4,输出answer的时间很长,是否可以把文本向量化的部分提前做好存储起来?,2023-04-10.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/50,GPU:4090 24G显存,4 -5,报错Use `repo_type` argument if needed.,2023-04-11.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/57,Traceback (most recent call last):,5 -6,无法打开gradio的页面,2023-04-11.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/58,$ python webui.py,6 -7,支持word,那word里面的图片正常显示吗?,2023-04-12.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/60,如题,刚刚从隔壁转过来的,想先了解下,7 -8,detectron2 is not installed. Cannot use the hi_res partitioning strategy. Falling back to partitioning with the fast strategy.,2023-04-12.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/63,能够正常的跑起来,在加载content文件夹中的文件时,每加载一个文件都会提示:,8 -9,cpu上运行webui,step3 asking时报错,2023-04-12.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/66,web运行,文件加载都正常,asking时报错,9 -10,建议弄一个插件系统,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/67,如题弄成stable-diffusion-webui那种能装插件,再开一个存储库给使用者或插件开发,存储或下载插件。,10 -11,请教加载模型出错!?,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/75,AttributeError: module 'transformers_modules.chatglm-6b.configuration_chatglm' has no attribute 'ChatGLMConfig 怎么解决呀,11 -12,从本地知识检索内容的时候,是否可以设置相似度阈值,小于这个阈值的内容不返回,即使会小于设置的VECTOR_SEARCH_TOP_K参数呢?谢谢大佬,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/76,比如 问一些 你好/你是谁 等一些跟本地知识库无关的问题,12 -13,如何改成多卡推理?,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/77,+1,13 -14,能否弄个懒人包,可以一键体验?,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/78,能否弄个懒人包,可以一键体验?,14 -15,连续问问题会导致崩溃,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/79,看上去不是爆内存的问题,连续问问题后,会出现如下报错,15 -16,AttributeError: 'NoneType' object has no attribute 'as_retriever',2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/86,"环境:windows 11, anaconda/python 3.8",16 -17,FileNotFoundError: Could not find module 'nvcuda.dll' (or one of its dependencies). Try using the full path with constructor syntax.,2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/87,请检查一下cuda或cudnn是否存在安装问题,17 -18,加载txt文件失败?,2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/89,![JppHrGOWFa](https://user-images.githubusercontent.com/109277248/232009383-bf7c46d1-a01e-4e0a-9de6-5b5ed3e36158.jpg),18 -19,NameError: name 'chatglm' is not defined,2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/90,"This share link expires in 72 hours. For free permanent hosting and GPU upgrades (NEW!), check out Spaces: https://huggingface.co/spaces",19 -20,打不开地址?,2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/91,报错数据如下:,20 -21,加载md文件出错,2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/98,运行 webui.py后能访问页面,上传一个md文件后,日志中有错误。等待后能加载完成,提示可以提问了,但提问没反应,日志中有错误。 具体日志如下。,21 -22,建议增加获取在线知识的能力,2023-04-15.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/101,建议增加获取在线知识的能力,22 -23,txt 未能成功加载,2023-04-15.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/103,hinese. Creating a new one with MEAN pooling.,23 -24,pdf加载失败,2023-04-15.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/105,e:\a.txt加载成功了,e:\a.pdf加载就失败,pdf文件里面前面几页是图片,后面都是文字,加载失败没有报更多错误,请问该怎么排查?,24 -25,一直停在文本加载处,2023-04-15.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/108,一直停在文本加载处,25 -26," File ""/root/.cache/huggingface/modules/transformers_modules/chatglm-6b/modeling_chatglm.py"", line 440, in forward new_tensor_shape = mixed_raw_layer.size()[:-1] + ( TypeError: torch.Size() takes an iterable of 'int' (item 2 is 'float')",2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/113,按照最新的代码,发现,26 -27,后续会提供前后端分离的功能吗?,2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/114,类似这种https://github.com/lm-sys/FastChat/tree/main/fastchat/serve,27 -28,安装依赖报错,2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/115,(test) C:\Users\linh\Desktop\langchain-ChatGLM-master>pip install -r requirements.txt,28 -29,问特定问题会出现爆显存,2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/117,正常提问没问题。,29 -30,Expecting value: line 1 column 1 (char 0),2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/118,运行后 第一步加载配置一直报错:,30 -31,embedding https://huggingface.co/GanymedeNil/text2vec-large-chinese/tree/main是免费的,效果比对openai的如何?,2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/119,-------------------------------------------------------------------------------,31 -32,这是什么错误,在Colab上运行的。,2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/120,libcuda.so.1: cannot open shared object file: No such file or directory,32 -33,只想用自己的lora微调后的模型进行对话,不想加载任何本地文档,该如何调整?,2023-04-18.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/121,能出一个单独的教程吗,33 -34,"租的gpu,Running on local URL: http://0.0.0.0:7860 To create a public link, set `share=True` in `launch()`. 浏览器上访问不了???",2023-04-18.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/122,(chatglm20230401) root@autodl-container-e82d11963c-10ece0d7:~/autodl-tmp/chatglm/langchain-ChatGLM-20230418# python3.9 webui.py,34 -35,本地部署中的报错请教,2023-04-18.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/124,"您好,在本地运行langchain-ChatGLM过程中,环境及依赖的包都已经满足条件,但是运行webui.py,报错如下(运行cli_demo.py报错类似),请问是哪里出了错呢?盼望您的回复,谢谢!",35 -36,报错。The dtype of attention mask (torch.int64) is not bool,2023-04-18.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/131,The dtype of attention mask (torch.int64) is not bool,36 -37,[求助] pip install -r requirements.txt 的时候出现以下报错。。。有大佬帮忙看看怎么搞么,下的release里面的包,2023-04-18.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/134,$ pip install -r requirements.txt,37 -38,如何提升根据问题搜索到对应知识的准确率,2023-04-19.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/136,外链知识库最大的问题在于问题是短文本,知识是中长文本。如何根据问题精准的搜索到对应的知识是个最大的问题。这类本地化项目不像百度,由无数的网页,基本上每个问题都可以找到对应的页面。,38 -39,是否可以增加向量召回的阈值设定,有些召回内容相关性太低,导致模型胡言乱语,2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/140,如题,39 -40,输入长度问题,2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/141,感谢作者支持ptuning微调模型。,40 -41,已有部署好的chatGLM-6b,如何通过接口接入?,2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/144,已有部署好的chatGLM-6b,如何通过接口接入,而不是重新加载一个模型;,41 -42,执行web_demo.py后,显示Killed,就退出了,是不是配置不足呢?,2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/146,![图片](https://user-images.githubusercontent.com/26102866/233256425-c7aab999-11d7-4de9-867b-23ef18d519e4.png),42 -43,执行python cli_demo1.py,2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/147,Traceback (most recent call last):,43 -44,报错:ImportError: cannot import name 'GENERATION_CONFIG_NAME' from 'transformers.utils',2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/149,(mychatGLM) PS D:\Users\admin3\zrh\langchain-ChatGLM> python cli_demo.py,44 -45,上传文件并加载知识库时,会不停地出现临时文件,2023-04-21.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/153,环境:ubuntu 18.04,45 -46,向知识库中添加文件后点击”上传文件并加载知识库“后Segmentation fault报错。,2023-04-23.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/161,运行服务后的提示如下:,46 -47,langchain-serve 集成,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/162,Hey 我是来自 [langchain-serve](https://github.com/jina-ai/langchain-serve) 的dev!,47 -48,大佬们,wsl的ubuntu怎么配置用cuda加速,装了运行后发现是cpu在跑,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/164,大佬们,wsl的ubuntu怎么配置用cuda加速,装了运行后发现是cpu在跑,48 -49,在github codespaces docker运行出错,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/165,docker run -d --restart=always --name chatglm -p 7860:7860 -v /www/wwwroot/code/langchain-ChatGLM:/chatGLM chatglm,49 -50,有计划接入Moss模型嘛,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/166,后续会开展测试,目前主要在优化langchain部分效果,如果有兴趣也欢迎提PR,50 -51,怎么实现 API 部署?,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/168,利用 fastapi 实现 API 部署方式,具体怎么实现,有方法说明吗?,51 -52, 'NoneType' object has no attribute 'message_types_by_name'报错,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/169,_HISTOGRAMPROTO = DESCRIPTOR.message_types_by_name['HistogramProto'],52 -53,能否指定自己训练的text2vector模型?,2023-04-25.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/172,请问大佬:,53 -54,关于项目支持的模型以及quantization_bit潜在的影响的问题,2023-04-26.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/176,作者您好~,54 -55,运行python3.9 api.py WARNING: You must pass the application as an import string to enable 'reload' or 'workers'.,2023-04-26.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/179,api.py文件最下面改成这样试试:,55 -56,ValidationError: 1 validation error for HuggingFaceEmbeddings model_kwargs extra fields not permitted (type=value_error.extra),2023-04-26.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/180,ValidationError: 1 validation error for HuggingFaceEmbeddings,56 -57,如果没有检索到相关性比较高的,回答“我不知道”,2023-04-26.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/181,如果通过设计system_template,让模型在搜索到的文档都不太相关的情况下回答“我不知道”,57 -58,请问如果不能联网,6B之类的文件从本地上传需要放到哪里,2023-04-26.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/182,感谢大佬的项目,很有启发~,58 -59,知识库问答--输入新的知识库名称是中文的话,会报error,2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/184,知识库问答--输入新的知识库名称是中文的话,会报error,选择要加载的知识库那里也不显示之前添加的知识库,59 -60,现在能通过问题匹配的相似度值,来直接返回文档中的文段,而不经过模型吗?因为有些答案在文档中,模型自己回答,不能回答文档中的答案,2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/186,现在能通过问题匹配的相似度值,来直接返回文档中的文段,而不经过模型吗?因为有些答案在文档中,模型自己回答,不能回答文档中的答案。也就是说,提供向量检索回答+模型回答相结合的策略。如果相似度值高于一定数值,直接返回文档中的文本,没有高于就返回模型的回答或者不知道,60 -61,"TypeError: The type of ChatGLM.callback_manager differs from the new default value; if you wish to change the type of this field, please use a type annotation",2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/188,"Mac 运行 python3 ./webui.py 报 TypeError: The type of ChatGLM.callback_manager differs from the new default value; if you wish to change the type of this field, please use a type annotation",61 -62,Not Enough Memory,2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/190,"运行命令行程序python cli_demo.py, 已经成功加载pdf文件, 报“DefaultCPUAllocator: not enough memory: you tried to allocate 458288380900 bytes”错误,请问哪里可以配置default memory",62 -63,参与开发问题,2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/191,1.是否需要进专门的开发群,63 -64,对话框中代码片段格式需改进,2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/192,最好能改进下输出代码片段的格式,目前输出的格式还不友好。,64 -65,请问未来有可能支持belle吗,2023-04-28.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/195,如题,谢谢大佬,65 -66,TypeError: cannot unpack non-iterable NoneType object,2023-04-28.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/200,"When i tried to change the knowledge vector store through `init_knowledge_vector_store`, the error `TypeError: cannot unpack non-iterable NoneType object` came out.",66 -67,生成结果,2023-04-28.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/202,你好,想问一下langchain+chatglm-6B,找到相似匹配的prompt,是直接返回prompt对应的答案信息,还是chatglm-6B在此基础上自己优化答案?,67 -68,在win、ubuntu下都出现这个错误:attributeerror: 't5forconditionalgeneration' object has no attribute 'stream_chat',2023-04-29.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/207,在win、ubuntu。下载完模型后,没办法修改代码以执行本地模型,每次都要重新输入路径; LLM 模型、Embedding 模型支持也都在官网下的,在其他项目(wenda)下可以使用,68 -69,[FEATURE] knowledge_based_chatglm.py: renamed or missing?,2023-04-30.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/210,"Not found. Was it renamed? Or, is it missing? How can I get it?",69 -70,sudo apt-get install -y nvidia-container-toolkit-base执行报错,2023-05-01.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/211,**问题描述 / Problem Description**,70 -71,效果不佳几乎答不上来,2023-05-01.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/212,提供了50条问答的docx文件,71 -72,有没有可能新增一个基于chatglm api调用的方式构建langchain,2023-05-02.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/218,我有两台8G GPU/40G内存的服务器,一个台做成了chatglm的api ;想基于另外一台服务器部署langchain;网上好像没有类似的代码。,72 -73,电脑是intel的集成显卡; 运行时告知我找不到nvcuda.dll,模型无法运行,2023-05-02.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/219,您好,我的电脑是intel的集成显卡,不过CPU是i5-11400 @ 2.60GHz ,内存64G;,73 -74,根据langchain官方的文档和使用模式,是否可以改Faiss为Elasticsearch?会需要做哪些额外调整?求解,2023-05-03.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/221,本人新手小白,由于业务模式的原因(有一些自己的场景和优化),希望利用Elasticsearch做这个体系内部的检索机制,不知道是否可以替换,同时,还会涉及到哪些地方的改动?或者说可能会有哪些其他影响,希望作者和大佬们不吝赐教!,74 -75,请问未来有可能支持t5吗,2023-05-04.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/224,请问可能支持基於t5的模型吗?,75 -76,[BUG] 内存溢出 / torch.cuda.OutOfMemoryError:,2023-05-04.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/229,**问题描述 / Problem Description**,76 -77,报错 No module named 'chatglm_llm',2023-05-04.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/230,明明已经安装了包,却在python里吊不出来,77 -78,能出一个api部署的描述文档吗,2023-05-04.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/233,**功能描述 / Feature Description**,78 -79,使用docs/API.md 出错,2023-05-04.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/234,使用API.md文档2种方法,出错,79 -80,加载pdf文档报错?,2023-05-05.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/238,ew one with MEAN pooling.,80 -81,上传的本地知识文件后再次上传不能显示,只显示成功了一个,别的上传成功后再次刷新就没了,2023-05-05.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/239,您好,项目有很大启发,感谢~,81 -82,创建了新的虚拟环境,安装了相关包,并且自动下载了相关的模型,但是仍旧出现:OSError: Unable to load weights from pytorch checkpoint file for,2023-05-05.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/240,![78ac8e663fdc312d0e9d78da95925c4](https://user-images.githubusercontent.com/34124260/236378728-9ea4424f-0f7f-4013-9d33-820b723de321.png),82 -83,[BUG] 数据加载不进来,2023-05-05.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/243,使用的.txt格式,utf-8编码,报以下错误,83 -84,不能读取pdf,2023-05-05.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/244,请问是webui还是cli_demo,84 -85,本地txt文件有500M,加载的时候很慢,如何提高速度?,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/251,![yayRzxSYHP](https://user-images.githubusercontent.com/109277248/236592902-f5ab338d-c1e9-43dc-ae16-9df2cd3c1378.jpg),85 -86,[BUG] gradio上传知识库后刷新之后 知识库就不见了 只有重启才能看到之前的上传的知识库,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/253,gradio上传知识库后刷新之后 知识库就不见了 只有重启才能看到之前的上传的知识库,86 -87,[FEATURE] 可以支持 OpenAI 的模型嘛?比如 GPT-3、GPT-3.5、GPT-4;embedding 增加 text-embedding-ada-002,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/254,**功能描述 / Feature Description**,87 -88,[FEATURE] 能否增加对于milvus向量数据库的支持 / Concise description of the feature,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/256,**功能描述 / Feature Description**,88 -89,CPU和GPU上跑,除了速度有区别,准确率效果回答上有区别吗?,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/259,理论上没有区别,89 -90,m1,请问在生成回答时怎么看是否使用了mps or cpu?,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/260,m1,请问在生成回答时怎么看是否使用了mps or cpu?,90 -91,知识库一刷新就没了,2023-05-07.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/263,知识库上传后刷新就没了,91 -92,本地部署报没有模型,2023-05-07.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/267,建议在下载llm和embedding模型至本地后在configs/model_config中写入模型本地存储路径后再运行,92 -93,[BUG] python3: can't open file 'webui.py': [Errno 2] No such file or directory,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/269,**问题描述 / Problem Description**,93 -94,模块缺失提示,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/271,因为已有自己使用的docker环境,直接启动webui.py,提示,94 -95,"运行api.py后,执行curl -X POST ""http://127.0.0.1:7861"" 报错?",2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/272,"执行curl -X POST ""http://127.0.0.1:7861"" \ -H 'Content-Type: application/json' \ -d '{""prompt"": ""你好"", ""history"": []}',报错怎么解决",95 -96,[BUG] colab安装requirements提示protobuf版本问题?,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/273,pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.,96 -97,请问项目里面向量相似度使用了什么方法计算呀?,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/275,基本按照langchain里的FAISS.similarity_search_with_score_by_vector实现,97 -98,[BUG] 安装detectron2后,pdf无法加载,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/276,**问题描述 / Problem Description**,98 -99,[BUG] 使用ChatYuan-V2模型无法流式输出,会报错,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/277,一方面好像是ChatYuan本身不支持stream_chat,有人在clueai那边提了issue他们说还没开发,所以估计这个attribute调不起来;但是另一方面看报错好像是T5模型本身就不是decoder-only模型,所以不能流式输出吧(个人理解),99 -100,[BUG] 无法加载text2vec模型,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/278,**问题描述 / Problem Description**,100 -101,请问能否增加网络搜索功能,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/281,请问能否增加网络搜索功能,101 -102,[FEATURE] 结构化数据sql、excel、csv啥时会支持呐。,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/283,**功能描述 / Feature Description**,102 -103,TypeError: ChatGLM._call() got an unexpected keyword argument 'stop',2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/284,No sentence-transformers model found with name D:\DevProject\langchain-ChatGLM\GanymedeNil\text2vec-large-chinese. Creating a new one with MEAN pooling.,103 -104,关于api.py的一些bug和设计逻辑问题?,2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/285,首先冒昧的问一下,这个api.py,开发者大佬们是在自己电脑上测试后确实没问题吗?,104 -105,有没有租用的算力平台上,运行api.py后,浏览器http://localhost:7861/报错,2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/287,是不是租用的gpu平台上都会出现这个问题???,105 -106,请问一下项目中有用到文档段落切割方法吗?,2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/288,text_load中的文档切割方法用上了吗?在代码中看好像没有用到?,106 -107,"报错 raise ValueError(f""Knowledge base {knowledge_base_id} not found"") ValueError: Knowledge base ./vector_store not found",2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/289,"File ""/root/autodl-tmp/chatglm/langchain-ChatGLM-master/api.py"", line 183, in chat",107 -108,能接入vicuna模型吗,2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/290,目前本地已经有了vicuna模型能直接接入吗?,108 -109,[BUG] 提问公式相关问题大概率爆显存,2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/291,**问题描述 / Problem Description**,109 -110,安装pycocotools失败,找了好多方法都不能解决。,2023-05-10.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/292,**问题描述 / Problem Description**,110 -111,使用requirements安装,PyTorch安装的是CPU版本,2023-05-10.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/294,如题目,使用requirements安装,PyTorch安装的是CPU版本,运行程序的时候,也是使用CPU在工作。,111 -112,能不能给一个毛坯服务器的部署教程,2023-05-10.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/298,“开发部署”你当成服务器的部署教程用就行了。,112 -113, Error(s) in loading state_dict for ChatGLMForConditionalGeneration:,2023-05-10.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/299,运行中出现的问题,7860的端口页面显示不出来,求助。,113 -114,ChatYuan-large-v2模型加载失败,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/300,**实际结果 / Actual Result**,114 -115,新增摘要功能,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/303,你好,后续会考虑新增对长文本信息进行推理和语音理解功能吗?比如生成摘要,115 -116,[BUG] pip install -r requirements.txt 出错,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/304,pip install langchain -i https://pypi.org/simple,116 -117,[BUG] 上传知识库文件报错,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/305,![19621e29eaa547d01213bee53d81e6a](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/84606552/7f6ceb46-e494-4b0e-939c-23b585a6d9d8),117 -118,[BUG] AssertionError: Component with id 41 not a valid input component.,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/306,**问题描述 / Problem Description**,118 -119,[BUG] CUDA out of memory with container deployment,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/310,**问题描述 / Problem Description**,119 -120,[FEATURE] 增加微调训练功能,2023-05-11.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/311,**功能描述 / Feature Description**,120 -121,如何使用多卡部署,多个gpu,2023-05-11.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/315,"机器上有多个gpu,如何全使用了",121 -122,请问这个知识库问答,和chatglm的关系是什么,2023-05-11.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/319,这个知识库问答,哪部分关联到了chatglm,是不是没有这个chatglm,知识库问答也可单单拎出来,122 -123,[BUG] 运行的时候报错ImportError: libcudnn.so.8: cannot open shared object file: No such file or directory,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/324,**问题描述 / Problem Description**raceback (most recent call last):,123 -124,webui启动成功,但有报错,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/325,**问题描述 / Problem Description**,124 -125,切换MOSS的时候报错,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/327,danshi但是发布的源码中,,125 -126,vicuna模型是否能接入?,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/328,您好!关于MOSS模型和vicuna模型,都是AutoModelForCausalLM来加载模型的,但是稍作更改(模型路径这些)会报这个错误。这个错误的造成是什么,126 -127,你好,请问一下在阿里云CPU服务器上跑可以吗?可以的话比较理想的cpu配置是什么?,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/330,你好,请问一下在阿里云CPU服务器上跑可以吗?可以的话比较理想的cpu配置是什么?,127 -128,你好,请问8核32g的CPU可以跑多轮对话吗?,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/331,什么样的cpu配置比较好呢?我目前想部署CPU下的多轮对话?,128 -129,[BUG] 聊天内容输入超过10000个字符系统出现错误,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/332,聊天内容输入超过10000个字符系统出现错误,如下图所示:,129 -130,能增加API的多用户访问接口部署吗?,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/333,默认部署程序仅支持单用户访问,多用户则需要排队访问。测试过相关的几个Github多用户工程,但是其中一些仍然不满足要求。本节将系统介绍如何实现多用户同时访问ChatGLM的部署接口,包括http、websocket(流式输出,stream)和web页面等方式,主要目录如下所示。,130 -131,多卡部署,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/334,用单机多卡或多机多卡,fastapi部署模型,怎样提高并发,131 -132,WEBUI能否指定知识库目录?,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/335,**功能描述 / Feature Description**,132 -133,[BUG] Cannot read properties of undefined (reading 'error'),2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/336,**问题描述 / Problem Description**,133 -134,[BUG] 1 validation error for HuggingFaceEmbeddings model_kwargs extra fields not permitted.,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/337,模型加载到 100% 后出现问题:,134 -135,上传知识库需要重启能不能修复一下,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/338,挺严重的这个问题,135 -136,[BUG] 4块v100卡爆显存,在LLM会话模式也一样,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/339,**问题描述 / Problem Description**,136 -137,针对上传的文件配置不同的TextSpliter,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/341,1. 目前的ChineseTextSpliter切分对英文尤其是代码文件不友好,而且限制固定长度;导致对话结果不如人意,137 -138,[FEATURE] 未来可增加Bloom系列模型吗?根据甲骨易的测试,这系列中文评测效果不错,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/346,**功能描述 / Feature Description**,138 -139,[BUG] v0.1.12打包镜像后启动webui.py失败 / Concise description of the issue,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/347,**问题描述 / Problem Description**,139 -140,切换MOSS模型时报错,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/349,昨天问了下,说是transformers版本不对,需要4.30.0,发现没有这个版本,今天更新到4.29.1,依旧报错,错误如下,140 -141,[BUG] pdf文档加载失败,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/350,**问题描述 / Problem Description**,141 -142,建议可以在后期增强一波注释,这样也有助于更多人跟进提PR,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/351,知道作者和团队在疯狂更新审查代码,只是建议后续稳定后可以把核心代码进行一些注释的补充,从而能帮助更多人了解各个模块作者的思路从而提出更好的优化。,142 -143,[FEATURE] MOSS 量化版支援,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/353,**功能描述 / Feature Description**,143 -144,[BUG] moss模型无法加载,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/356,**问题描述 / Problem Description**,144 -145,[BUG] load_doc_qa.py 中的 load_file 函数有bug,2023-05-14.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/358,原函数为:,145 -146,[FEATURE] API模式,知识库加载优化,2023-05-14.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/359,如题,当前版本,每次调用本地知识库接口,都将加载一次知识库,是否有更好的方式?,146 -147,运行Python api.py脚本后端部署后,怎么使用curl命令调用?,2023-05-15.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/361,也就是说,我现在想做个对话机器人,想和公司的前后端联调?怎么与前后端相互调用呢?可私信,有偿解答!!!,147 -148,上传知识库需要重启能不能修复一下,2023-05-15.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/363,上传知识库需要重启能不能修复一下,148 -149,[BUG] pip install -r requirements.txt -i https://pypi.tuna.tsinghua.edu.cn/simple,2023-05-15.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/364,我的python是3.8.5的,149 -150,pip install gradio 报错,2023-05-15.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/367,大佬帮我一下,150 -151,[BUG] pip install gradio 一直卡不动,2023-05-15.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/369,![aba82742dd9d4d242181662eb5027a7](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/84606552/cd9600d9-f6e7-46b7-b1be-30ed8b99f76b),151 -152,[BUG] 简洁阐述问题 / Concise description of the issue,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/370,初次加载本地知识库成功,但提问后,就无法重写加载本地知识库,152 -153,[FEATURE] 简洁阐述功能 / Concise description of the feature,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/371,**功能描述 / Feature Description**,153 -154,在windows上,模型文件默认会安装到哪,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/372,-------------------------------------------------------------------------------,154 -155,[FEATURE] 兼顾对话管理,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/374,如何在知识库检索的情况下,兼顾对话管理?,155 -156,llm device: cpu embedding device: cpu,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/376,**问题描述 / Problem Description**,156 -157,[FEATURE] 简洁阐述功能 /文本文件的知识点之间使用什么分隔符可以分割?,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/377,**功能描述 / Feature Description**,157 -158,[BUG] 上传文件失败:PermissionError: [WinError 32] 另一个程序正在使用此文件,进程无法访问。,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/379,**问题描述 / Problem Description**,158 -159,[BUG] 执行python api.py 报错,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/383,错误信息,159 -160,model_kwargs extra fields not permitted (type=value_error.extra),2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/384,"大家好,请问这个有遇到的么,?",160 -161,[BUG] 简洁阐述问题 / Concise description of the issue,2023-05-17.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/385,执行的时候出现了ls1 = [ls[0]],161 -162,[FEATURE] 性能优化,2023-05-17.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/388,**功能描述 / Feature Description**,162 -163,"[BUG] Moss模型问答,RuntimeError: probability tensor contains either inf, nan or element < 0",2023-05-17.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/390,**问题描述 / Problem Description**,163 -164,有没有人知道v100GPU的32G显存,会报错吗?支持V100GPU吗?,2023-05-17.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/392,**问题描述 / Problem Description**,164 -165,针对于编码问题比如'gbk' codec can't encode character '\xab' in position 14: illegal multibyte sequence粗浅的解决方法,2023-05-17.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/397,**功能描述 / Feature Description**,165 -166,Could not import sentence_transformers python package. Please install it with `pip install sentence_transformers`.,2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/400,**问题描述 / Problem Description**,166 -167,支持模型问答与检索问答,2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/401,不同的query,根据意图不一致,回答也应该不一样。,167 -168,文本分割的时候,能不能按照txt文件的每行进行分割,也就是按照换行符号\n进行分割???,2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/403,下面的代码应该怎么修改?,168 -169,local_doc_qa/local_doc_chat 接口响应是串行,2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/405,**问题描述 / Problem Description**,169 -170,"为什么找到出处了,但是还是无法回答该问题?",2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/406,![图片](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/3349611/1fc81d61-2409-4330-9065-fdda1a27c86a),170 -171,"请问下:知识库测试中的:添加单条内容,如果换成文本导入是是怎样的格式?我发现添加单条内容测试效果很好.",2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/412,"我发现在知识库测试中`添加单条内容`,并且勾选`禁止内容分句入库`,即使 `不开启上下文关联`的测试效果都非常好.",171 -172,[BUG] 无法配置知识库,2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/413,**问题描述 / Problem Description**,172 -173,[BUG] 部署在阿里PAI平台的EAS上访问页面是白屏,2023-05-19.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/414,**问题描述 / Problem Description**,173 -174,API部署后调用/local_doc_qa/local_doc_chat 返回Knowledge base samples not found,2023-05-19.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/416,入参,174 -175,[FEATURE] 上传word另存为的txt文件报 'ascii' codec can't decode byte 0xb9 in position 6: ordinal not in range(128),2023-05-20.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/421,上传word另存为的txt文件报,175 -176,创建保存的知识库刷新后没有出来,这个知识库是永久保存的吗?可以连外部的 向量知识库吗?,2023-05-21.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/422,创建保存的知识库刷新后没有出来,这个知识库是永久保存的吗?可以连外部的 向量知识库吗?,176 -177,[BUG] 用colab运行,无法加载模型,报错:'NoneType' object has no attribute 'message_types_by_name',2023-05-21.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/423,**问题描述 / Problem Description**,177 -178,请问是否需要用到向量数据库?以及什么时候需要用到向量数据库?,2023-05-21.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/424,目前用的是 text2vec , 请问是否需要用到向量数据库?以及什么时候需要用到向量数据库?,178 -179,huggingface模型引用问题,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/427,它最近似乎变成了一个Error?,179 -180,你好,加载本地txt文件出现这个killed错误,TXT文件有100M左右大小。原因是?谢谢。,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/429,"",180 -181,想请问一下,关于对本地知识的管理是如何管理?例如:通过http API接口添加数据 或者 删除某条数据,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/430,例如:通过http API接口添加、删除、修改 某条数据。,181 -182,[FEATURE] 双栏pdf识别问题,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/432,试了一下模型,感觉对单栏pdf识别的准确性较高,但是由于使用的基本是ocr的技术,对一些双栏pdf论文识别出来有很多问题,请问有什么办法改善吗?,182 -183,部署启动小问题,小弟初学求大佬解答,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/433,1.python loader/image_loader.py时,提示ModuleNotFoundError: No module named 'configs',但是跑python webui.py还是还能跑,183 -184,能否支持检测到目录下文档有增加而去增量加载文档,不影响前台对话,其实就是支持读写分离。如果能支持查询哪些文档向量化了,删除过时文档等就更好了,谢谢。,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/434,**功能描述 / Feature Description**,184 -185,[BUG] 简洁阐述问题 / windows 下cuda错误,请用https://github.com/Keith-Hon/bitsandbytes-windows.git,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/435,pip install git+https://github.com/Keith-Hon/bitsandbytes-windows.git,185 -186,"[BUG] from commit 33bbb47, Required library version not found: libbitsandbytes_cuda121_nocublaslt.so. Maybe you need to compile it from source?",2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/438,**问题描述 / Problem Description**,186 -187,[BUG] 简洁阐述问题 / Concise description of the issue上传60m的txt文件报错,显示超时,请问这个能上传的文件大小有限制吗,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/439,"ERROR 2023-05-23 11:13:09,627-1d: Timeout reached while detecting encoding for ./docs/GLM模型格式数据.txt",187 -188,[BUG] TypeError: issubclass() arg 1 must be a class,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/440,**问题描述**,188 -189,"执行python3 webui.py后,一直提示”模型未成功加载,请到页面左上角""模型配置""选项卡中重新选择后点击""加载模型""按钮“",2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/441,**问题描述 / Problem Description**,189 -190,是否能提供网页文档得导入支持,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/444,现在很多都是在线文档作为协作得工具,所以通过URL导入在线文档需求更大,190 -191,[BUG] history 索引问题,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/445,在比较对话框的history和模型chat function 中的history时, 发现并不匹配,在传入 llm._call 时,history用的索引是不是有点问题,导致上一轮对话的内容并不输入给模型。,191 -192,[BUG] moss_llm没有实现,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/447,有些方法没支持,如history_len,192 -193,请问langchain-ChatGLM如何删除一条本地知识库的数据?,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/448,例如:用户刚刚提交了一条错误的数据到本地知识库中了,现在如何在本地知识库从找到,并且对此删除。,193 -194,[BUG] 简洁阐述问题 / UnboundLocalError: local variable 'resp' referenced before assignment,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/450,"在最新一版的代码中, 运行api.py 出现了以上错误(UnboundLocalError: local variable 'resp' referenced before assignment), 通过debug的方式观察到local_doc_qa.llm.generatorAnswer(prompt=question, history=history,streaming=True)可能不返回任何值。",194 -195,请问有没有 PROMPT_TEMPLATE 能让模型不回答敏感问题,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/452,## PROMPT_TEMPLATE问题,195 -196,[BUG] 测试环境 Python 版本有误,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/456,**问题描述 / Problem Description**,196 -197,[BUG] webui 部署后样式不正确,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/458,**问题描述 / Problem Description**,197 -198,配置默认LLM模型的问题,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/459,**问题描述 / Problem Description**,198 -199,[FEATURE]是时候更新一下autoDL的镜像了,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/460,如题,跑了下autoDL的镜像,发现是4.27号的,git pull新版本的代码功能+老的依赖环境,各种奇奇怪怪的问题。,199 -200,[BUG] tag:0.1.13 以cpu模式下,想使用本地模型无法跑起来,各种路径参数问题,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/462,-------------------------------------------------------------------------------,200 -201,[BUG] 有没有同学遇到过这个错!!!加载本地txt文件出现这个killed错误,TXT文件有100M左右大小。,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/463,运行cli_demo.py。是本地的txt文件太大了吗?100M左右。,201 -202,API版本能否提供WEBSOCKET的流式接口,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/464,webui 版本中,采用了WS的流式输出,整体感知反应很快,202 -203,[BUG] 安装bug记录,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/465,按照[install文档](https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/docs/INSTALL.md)安装的,,203 -204,VUE的pnmp i执行失败的修复-用npm i命令即可,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/466,感谢作者!非常棒的应用,用的很开心。,204 -205,请教个问题,有没有人知道cuda11.4是否支持???,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/467,请教个问题,有没有人知道cuda11.4是否支持???,205 -206,请问有实现多轮问答中基于问题的搜索上下文关联么,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/468,在基于知识库的多轮问答中,第一个问题讲述了一个主题,后续的问题描述没有包含这个主题的关键词,但又存在上下文的关联。如果用后续问题去搜索知识库有可能会搜索出无关的信息,从而导致大模型无法正确回答问题。请问这个项目要考虑这种情况吗?,206 -207,[BUG] 内存不足的问题,2023-05-26.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/470,我用了本地的chatglm-6b-int4模型,然后显示了内存不足(win10+32G内存+1080ti11G),一般需要多少内存才足够?这个bug应该如何解决?,207 -208,[BUG] 纯内网环境安装pycocotools失败,2023-05-26.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/472,**问题描述 / Problem Description**,208 -209,[BUG] webui.py 重新加载模型会导致 KeyError,2023-05-26.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/473,**问题描述 / Problem Description**,209 -210,chatyuan无法使用,2023-05-26.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/475,**问题描述 / Problem Description**,210 -211,[BUG] 文本分割模型AliTextSplitter存在bug,会把“.”作为分割符,2023-05-26.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/476,"阿里达摩院的语义分割模型存在bug,默认会把"".”作为分割符进行分割而不管上下文语义。是否还有其他分割符则未知。建议的修改方案:把“.”统一替换为其他字符,分割后再替换回来。或者添加其他分割模型。",211 -212,[BUG] RuntimeError: Error in faiss::FileIOReader::FileIOReader(const char*) a,2023-05-27.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/479,**问题描述 / Problem Description**,212 -213,[FEATURE] 安装,为什么conda create要额外指定路径 用-p ,而不是默认的/envs下面,2023-05-28.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/481,##**功能描述 / Feature Description**,213 -214,[小白求助] 通过Anaconda执行webui.py后,无法打开web链接,2023-05-28.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/485,在执行webui.py命令后,http://0.0.0.0:7860复制到浏览器后无法打开,显示“无法访问此网站”。,214 -215,[BUG] 使用 p-tuningv2后的模型,重新加载报错,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/486,把p-tunningv2训练完后的相关文件放到了p-tunningv2文件夹下,勾选使用p-tuningv2点重新加载模型,控制台输错错误信息:,215 -216,[小白求助] 服务器上执行webui.py后,在本地无法打开web链接,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/487,此项目执行在xxx.xx.xxx.xxx服务器上,我在webui.py上的代码为 (demo,216 -217,[FEATURE] 能不能支持VisualGLM-6B,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/488,**功能描述 / Feature Description**,217 -218,你好,问一下各位,后端api部署的时候,支持多用户同时问答吗???,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/489,支持多用户的话,最多支持多少用户问答?根据硬件而定吧?,218 -219,V100GPU显存占满,而利用率却为0,这是为什么?,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/491,"",219 -220,[求助] 如果在公司内部搭建产品知识库,使用INT-4模型,200人规模需要配置多少显存的服务器?,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/492,如题,计划给公司搭一个在线知识库。,220 -221,你好,请教个问题,目前问答回复需要20秒左右,如何提高速度?V10032G服务器。,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/493,**问题描述 / Problem Description**,221 -222,[FEATURE] 如何实现只匹配下文,而不要上文的结果,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/494,在构建自己的知识库时,主要采用问答对的形式,那么也就是我需要的回答是在我的问题下面的内容,但是目前设置了chunk_size的值以后匹配的是上下文的内容,但我实际并不需要上文的。为了实现更完整的展示下面的答案,我只能调大chunk_size的值,但实际上上文的一半内容都是我不需要的。也就是扔了一半没用的东西给prompt,在faiss.py中我也没找到这块的一些描述,请问该如何进行修改呢?,222 -223,你好,问一下,我调用api.py部署,为什么用ip加端口可以使用postman调用,而改为域名使用postman无法调用?,2023-05-30.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/497,![5ufBSWxLyF](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/109277248/70e2fbac-5699-48d0-b0d1-3dc84fd042c2),223 -224,调用api.py中的stream_chat,返回source_documents中出现中文乱码。,2023-05-30.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/498,-------------------------------------------------------------------------------,224 -225,[BUG] 捉个虫,api.py中的stream_chat解析json问题,2023-05-30.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/501,**问题描述 / Problem Description**,225 -226,windows本地部署遇到了omp错误,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/502,**问题描述 / Problem Description**,226 -227,"[BUG] bug14 ,""POST /local_doc_qa/upload_file HTTP/1.1"" 422 Unprocessable Entity",2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/503,上传的文件报错,返回错误,api.py,227 -228,你好,请教个问题,api.py部署的时候,如何改为多线程调用?谢谢,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/505,目前的api.py脚本不支持多线程,228 -229,你好,请教一下。api.py部署的时候,能不能提供给后端流失返回结果。,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/507,curl -X 'POST' \,229 -230,流式输出,流式接口,使用server-sent events技术。,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/508,想这样一样,https://blog.csdn.net/weixin_43228814/article/details/130063010,230 -231,计划增加流式输出功能吗?ChatGLM模型通过api方式调用响应时间慢怎么破,Fastapi流式接口来解惑,能快速提升响应速度,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/509,**问题描述 / Problem Description**,231 -232,[BUG] 知识库上传时发生ERROR (could not open xxx for reading: No such file or directory),2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/510,**问题描述 / Problem Description**,232 -233,api.py脚本打算增加SSE流式输出吗?,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/511,curl调用的时候可以检测第一个字,从而提升回复的体验,233 -234,[BUG] 使用tornado实现webSocket,可以多个客户端同时连接,并且实现流式回复,但是多个客户端同时使用,答案就很乱,是模型不支持多线程吗,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/512,import asyncio,234 -235,支持 chinese_alpaca_plus_lora 吗 基于llama的,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/514,支持 chinese_alpaca_plus_lora 吗 基于llama的,https://github.com/ymcui/Chinese-LLaMA-Alpaca这个项目的,235 -236,[BUG] 现在能读图片的pdf了,但是文字的pdf反而读不了了,什么情况???,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/515,**问题描述 / Problem Description**,236 -237,在推理的过程中卡住不动,进程无法正常结束,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/516,**问题描述 / Problem Description**,237 -238,curl调用的时候,从第二轮开始,curl如何传参可以实现多轮对话?,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/517,第一轮调用:,238 -239,建议添加api.py部署后的日志管理功能?,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/518,-------------------------------------------------------------------------------,239 -240,有大佬知道,怎么多线程部署api.py脚本吗?,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/519,api.py部署后,使用下面的请求,时间较慢,好像是单线程,如何改为多线程部署api.py:,240 -241,[BUG] 上传文件到知识库 任何格式与内容都永远失败,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/520,上传知识库的时候,传txt无法解析,就算是穿content/sample里的样例txt也无法解析,上传md、pdf等都无法加载,会持续性等待,等到了超过30分钟也不行。,241 -242,关于prompt_template的问题,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/521,请问这段prompt_template是什么意思,要怎么使用?可以给一个具体模板参考下吗?,242 -243,[BUG] 简洁阐述问题 / Concise description of the issue,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/522,**问题描述 / Problem Description**,243 -244,"中文分词句号处理(关于表达金额之间的""."")",2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/523,建议处理12.6亿元的这样的分词,最好别分成12 和6亿这样的,需要放到一起,244 -245,ImportError: cannot import name 'inference' from 'paddle',2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/526,在网上找了一圈,有说升级paddle的,我做了还是没有用,有说安装paddlepaddle的,我找了豆瓣的镜像源,但安装报错cannot detect archive format,245 -246,[BUG] webscoket 接口串行问题(/local_doc_qa/stream-chat/{knowledge_base_id}),2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/527,**问题描述 / Problem Description**,246 -247,[FEATURE] 刷新页面更新知识库列表,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/528,**功能描述以及改进方案**,247 -248,[BUG] 使用ptuning微调模型后,问答效果并不好,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/530,### 未调用ptuning,248 -249,[BUG] 多轮对话效果不佳,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/532,在进行多轮对话的时候,无论设置的history_len是多少,效果都不好。事实上我将其设置成了最大值10,但在对话中,仍然无法实现多轮对话:,249 -250,"RuntimeError: MPS backend out of memory (MPS allocated: 18.00 GB, other allocations: 4.87 MB, max allowed: 18.13 GB)",2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/533,**问题描述**,250 -251, 请大家重视这个issue!真正使用肯定是多用户并发问答,希望增加此功能!!!,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/534,这得看你有多少显卡,251 -252,在启动项目的时候如何使用到多张gpu啊?,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/535,**在启动项目的时候如何使用到多张gpu啊?**,252 -253, 使用流式输出的时候,curl调用的格式是什么?,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/536,"app.websocket(""/local_doc_qa/stream-chat/{knowledge_base_id}"")(stream_chat)中的knowledge_base_id应该填什么???",253 -254,使用本地 vicuna-7b模型启动错误,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/538,环境: ubuntu 22.04 cuda 12.1 没有安装nccl,使用rtx2080与m60显卡并行计算,254 -255,为什么会不调用GPU直接调用CPU呢,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/539,我的阿里云配置是16G显存,用默认代码跑webui.py时提示,255 -256,上传多个文件时会互相覆盖,2023-06-03.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/541,1、在同一个知识库中上传多个文件时会互相覆盖,无法结合多个文档的知识,有大佬知道怎么解决吗?,256 -257,[BUG] ‘gcc’不是内部或外部命令/LLM对话只能持续一轮,2023-06-03.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/542,No compiled kernel found.,257 -258,以API模式启动项目却没有知识库的接口列表?,2023-06-04.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/544,请问如何获取知识库的接口列表?如果没有需要自行编写的话,可不可以提供相关的获取方式,感谢,258 -259,程序以API模式启动的时候,如何才能让接口以stream模式被调用呢?,2023-06-05.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/546,作者您好,我在以API模式进行程序启动后,我发现接口响应时间很长,怎么样才能让接口以stream模式被调用呢?我想实现像webui模式的回答那样,259 -260,关于原文中表格转为文本后数据相关度问题。,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/547,原文中表格数据转换为文本,以 (X-Y:值;...) 的格式每一行组织成一句话,但这样做后发现相关度较低,效果很差,有何好的方案吗?,260 -261,启动后LLM和知识库问答模式均只有最后一轮记录,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/548,拉取最新代码,问答时,每次页面只显示最后一次问答记录,需要修改什么参数才可以保留历史记录?,261 -262,提供system message配置,以便于让回答不要超出知识库范围,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/549,**功能描述 / Feature Description**,262 -263,[BUG] 使用p-tunningv2报错,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/551,按照readme的指示把p-tunningv2训练完后的文件放到了p-tunningv2文件夹下,勾选使用p-tuningv2点重新加载模型,控制台提示错误信息:,263 -264,[BUG] 智障,这么多问题,也好意思放出来,浪费时间,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/553,。。。,264 -265,[FEATURE] 我看代码文件中有一个ali_text_splitter.py,为什么不用他这个文本分割器了?,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/554,我看代码文件中有一个ali_text_splitter.py,为什么不用他这个文本分割器了?,265 -266,加载文档函数报错,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/557,"def load_file(filepath, sentence_size=SENTENCE_SIZE):",266 -267,参考指引安装docker后,运行cli_demo.py,提示killed,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/558,root@b3d1bd08095c:/chatGLM# python3 cli_demo.py,267 -268,注意:如果安装错误,注意这两个包的版本 wandb==0.11.0 protobuf==3.18.3,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/559,Error1: 如果启动异常报错 `protobuf` 需要更新到 `protobuf==3.18.3 `,268 -269,知识库对长文的知识相关度匹配不太理想有何优化方向,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/563,我们可能录入一个文章有 1W 字,里面涉及这个文章主题的很多角度问题,我们针对他提问,他相关度匹配的内容和实际我们需要的答案相差很大怎么办。,269 -270,使用stream-chat函数进行流式输出的时候,能使用curl调用吗?,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/565,为什么下面这样调用会报错???,270 -271,有大佬实践过 并行 或者 多线程 的部署方案吗?,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/566,+1,271 -272,多线程部署遇到问题?,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/567,"",272 -273,[BUG] 用fastchat加载vicuna-13b模型进行知识库的问答有token的限制错误,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/569,当我开启fastchat的vicuna-13b的api服务,然后config那里配置好(api本地测试过可以返回结果),然后知识库加载好之后(知识库大概有1000多个文档,用chatGLM可以正常推理),进行问答时出现token超过限制,就问了一句hello;,273 -274,现在的添加知识库,文件多了总是报错,也不知道自己加载了哪些文件,报错后也不知道是全部失败还是一部分成功;希望能有个加载指定文件夹作为知识库的功能,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/574,**功能描述 / Feature Description**,274 -275,[BUG] moss模型本地加载报错,2023-06-08.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/577,moss模型本地加载报错:,275 -276,加载本地moss模型报错Can't instantiate abstract class MOSSLLM with abstract methods _history_len,2023-06-08.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/578,(vicuna) ps@ps[13:56:20]:/data/chat/langchain-ChatGLM2/langchain-ChatGLM-0.1.13$ python webui.py --model-dir local_models --model moss --no-remote-model,276 -277,[FEATURE] 能增加在前端页面控制prompt_template吗?或是能支持前端页面选择使用哪个prompt?,2023-06-08.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/579,目前只能在config里修改一个prompt,想在多个不同场景切换比较麻烦,277 -278,[BUG] streamlit ui的bug,在增加知识库时会报错,2023-06-08.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/580,**问题描述 / Problem Description**,278 -279,[FEATURE] webui/webui_st可以支持history吗?目前仅能一次对话,2023-06-08.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/581,试了下webui和webui_st都不支持历史对话啊,只能对话一次,不能默认开启所有history吗?,279 -280,启动python cli_demo.py --model chatglm-6b-int4-qe报错,2023-06-09.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/585,下载好模型,和相关依赖环境,之间运行`python cli_demo.py --model chatglm-6b-int4-qe`报错了:,280 -281,重新构建知识库报错,2023-06-09.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/586,**问题描述 / Problem Description**,281 -282,[FEATURE] 能否屏蔽paddle,我不需要OCR,效果差依赖环境还很复杂,2023-06-09.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/587,希望能不依赖paddle,282 -283,question :文档向量化这个可以自己手动实现么?,2023-06-09.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/589,现有公司级数据500G+,需要使用这个功能,请问如何手动实现这个向量化,然后并加载,283 -284,view前端能进行流式的返回吗??,2023-06-09.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/590,view前端能进行流式的返回吗??,284 -285,"[BUG] Load parallel cpu kernel failed, using default cpu kernel code",2023-06-11.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/594,**问题描述 / Problem Description**,285 -286,[BUG] 简洁阐述问题 / Concise description of the issue,2023-06-11.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/595,**问题描述 / Problem Description**,286 -287,我在上传本地知识库时提示KeyError: 'name'错误,本地知识库都是.txt文件,文件数量大约是2000+。,2023-06-12.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/597,"",287 -288,model_config.py中有vicuna-13b-hf模型的配置信息,但是好像还是不可用?,2023-06-12.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/600,@dongyihua543,288 -289,"ImportError: Using SOCKS proxy, but the 'socksio' package is not installed. Make sure to install httpx using `pip install httpx[socks]`.",2023-06-12.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/605,应该代理问题,但是尝试了好多方法都解决不了,,289 -290,[BUG] similarity_search_with_score_by_vector在找不到匹配的情况下出错,2023-06-12.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/607,在设置匹配阈值 VECTOR_SEARCH_SCORE_THRESHOLD 的情况下,vectorstore会返回空,此时上述处理函数会出错,290 -291,[FEATURE] 请问如何搭建英文知识库呢,2023-06-12.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/609,**功能描述 / Feature Description**,291 -292,谁有vicuna权重?llama转换之后的,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/611,**问题描述 / Problem Description**,292 -293,[FEATURE] API能实现上传文件夹的功能么?,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/612,用户懒得全选所有的文件,就想上传个文件夹,请问下API能实现这个功能么?,293 -294,请问在多卡部署后,上传单个文件作为知识库,用的是单卡在生成向量还是多卡?,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/614,目前我检测我本地多卡部署的,好像生成知识库向量的时候用的还是单卡,294 -295,[BUG] python webui.py提示非法指令,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/615,(/data/conda-langchain [root@chatglm langchain-ChatGLM]# python webui.py,295 -296,知识库文件跨行切分问题,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/616,我的知识库文件txt文件,是一行一条知识,用\n分行。,296 -297,[FEATURE] bing搜索问答有流式的API么?,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/617,web端是有这个bing搜索回答,但api接口没有发现,大佬能给个提示么?,297 -298,希望出一个macos m2的安装教程,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/620,mac m2安装,模型加载成功了,知识库文件也上传成功了,但是一问答就会报错,报错内容如下,298 -299,为【出处】提供高亮显示,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/621,具体出处里面,对相关的内容高亮显示,不包含前后文。,299 -300,[BUG] CPU运行cli_demo.py,不回答,hang住,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/622,没有GPU;32G内存的ubuntu机器。,300 -301,关于删除知识库里面的文档后,LLM知识库对话的时候还是会返回该被删除文档的内容,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/623,如题,在vue前端成功执行删除知识库里面文档A.txt后,未能也在faiss索引中也删除该文档,LLM还是会返回这个A.txt的内容,并且以A.txt为出处,未能达到删除的效果,301 -302,"[BUG] 调用知识库进行问答,显存会一直叠加",2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/625,"14G的显存,调用的chatglm-6b-int8模型,进行知识库问答时,最多问答四次就会爆显存了,观察了一下显存使用情况,每一次使用就会增加一次显存,请问这样是正常的吗?是否有什么配置需要开启可以解决这个问题?例如进行一次知识库问答清空上次问题的显存?",302 -303,[BUG] web页面 重新构建数据库 失败,导致 原来的上传的数据库都没了,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/626,web页面 重新构建数据库 失败,导致 原来的上传的数据库都没了,303 -304,在CPU上运行webui.py报错Tensor on device cpu is not on the expected device meta!,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/627,在CPU上运行python webui.py能启动,但最后有:RuntimeError: Tensor on device cpu is not on the expected device meta!,304 -305,"OSError: [WinError 1114] 动态链接库(DLL)初始化例程失败。 Error loading ""E:\xxx\envs\langchain\lib\site-packages\torch\lib\caffe2_nvrtc.dll"" or one of its dependencies.哪位大佬知道如何解决吗?",2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/629,**问题描述 / Problem Description**,305 -306,[BUG] WEBUI删除知识库文档,会导致知识库问答失败,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/632,如题,从知识库已有文件中选择要删除的文件,点击删除后,在问答框输入内容回车报错,306 -307,更新后的版本中,删除知识库中的文件,再提问出现error错误,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/634,针对更新版本,识别到一个问题,过程如下:,307 -308,我配置好了环境,想要实现本地知识库的问答?可是它返回给我的,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/637,没有总结,只有相关度的回复,但是我看演示里面表现的,回复是可以实现总结的,我去查询代码,308 -309,[BUG] NPM run dev can not successfully start the VUE frontend,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/638,**问题描述 / Problem Description**,309 -310,[BUG] 简洁阐述问题 / Concise description of the issue,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/639,**问题描述 / Problem Description**,310 -311,提一个模型加载的bug,我在截图中修复了,你们有空可以看一下。,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/642,![model_load_bug](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/59411575/4432adc4-ccdd-45d9-aafc-5f2d1963403b),311 -312,[求助]关于设置embedding model路径的问题,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/643,如题,我之前成功跑起来过一次,但因环境丢失重新配置 再运行webui就总是报错,312 -313,Lora微调后的模型可以直接使用吗,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/646,看model_config.py里是有USE_LORA这个参数的,但是在cli_demo.py和webui.py这两个里面都没有用到,实际测试下来模型没有微调的效果,想问问现在这个功能实现了吗,313 -314,write_check_file在tmp_files目录下生成的load_file.txt是否需要一直保留,占用空间很大,在建完索引后能否删除,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/647,**功能描述 / Feature Description**,314 -315,[BUG] /local_doc_qa/list_files?knowledge_base_id=test删除知识库bug,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/649,1.新建test知识库并上传文件(在vue前端完成并检查后端发现确实生成了test文件夹以及下面的content和vec_store,315 -316,[BUG] vue webui无法加载知识库,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/650,拉取了最新的代码,分别运行了后端api和前端web,点击知识库,始终只能显示simple,无法加载知识库,316 -317,不能本地加载moss模型吗?,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/652,手动下载模型设置local_model_path路径依旧提示缺少文件,该如何正确配置?,317 -318,macos m2 pro docker 安装失败,2023-06-17.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/654,macos m2 pro docker 安装失败,318 -319, [BUG] mac m1 pro 运行提示 zsh: segmentation fault,2023-06-17.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/655,运行: python webui.py,319 -320,安装 requirements 报错,2023-06-17.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/656,(langchainchatglm) D:\github\langchain-ChatGLM>pip install -r requirements.txt -i https://pypi.tuna.tsinghua.edu.cn/simple/,320 -321,[BUG] AssertionError,2023-06-17.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/658,**问题描述 / Problem Description**,321 -322,[FEATURE] 支持AMD win10 本地部署吗?,2023-06-18.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/660,**功能描述 / Feature Description**,322 +,title,file,url,detail,id +0,效果如何优化,2023-04-04.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/14,如图所示,将该项目的README.md和该项目结合后,回答效果并不理想,请问可以从哪些方面进行优化,0 +1,怎么让模型严格根据检索的数据进行回答,减少胡说八道的回答呢,2023-04-04.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/15,举个例子:,1 +2,"When I try to run the `python knowledge_based_chatglm.py`, I got this error in macOS(M1 Max, OS 13.2)",2023-04-07.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/32,```python,2 +3,萌新求教大佬怎么改成AMD显卡或者CPU?,2023-04-10.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/48,把.cuda()去掉就行,3 +4,输出answer的时间很长,是否可以把文本向量化的部分提前做好存储起来?,2023-04-10.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/50,GPU:4090 24G显存,4 +5,报错Use `repo_type` argument if needed.,2023-04-11.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/57,Traceback (most recent call last):,5 +6,无法打开gradio的页面,2023-04-11.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/58,$ python webui.py,6 +7,支持word,那word里面的图片正常显示吗?,2023-04-12.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/60,如题,刚刚从隔壁转过来的,想先了解下,7 +8,detectron2 is not installed. Cannot use the hi_res partitioning strategy. Falling back to partitioning with the fast strategy.,2023-04-12.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/63,能够正常的跑起来,在加载content文件夹中的文件时,每加载一个文件都会提示:,8 +9,cpu上运行webui,step3 asking时报错,2023-04-12.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/66,web运行,文件加载都正常,asking时报错,9 +10,建议弄一个插件系统,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/67,如题弄成stable-diffusion-webui那种能装插件,再开一个存储库给使用者或插件开发,存储或下载插件。,10 +11,请教加载模型出错!?,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/75,AttributeError: module 'transformers_modules.chatglm-6b.configuration_chatglm' has no attribute 'ChatGLMConfig 怎么解决呀,11 +12,从本地知识检索内容的时候,是否可以设置相似度阈值,小于这个阈值的内容不返回,即使会小于设置的VECTOR_SEARCH_TOP_K参数呢?谢谢大佬,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/76,比如 问一些 你好/你是谁 等一些跟本地知识库无关的问题,12 +13,如何改成多卡推理?,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/77,+1,13 +14,能否弄个懒人包,可以一键体验?,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/78,能否弄个懒人包,可以一键体验?,14 +15,连续问问题会导致崩溃,2023-04-13.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/79,看上去不是爆内存的问题,连续问问题后,会出现如下报错,15 +16,AttributeError: 'NoneType' object has no attribute 'as_retriever',2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/86,"环境:windows 11, anaconda/python 3.8",16 +17,FileNotFoundError: Could not find module 'nvcuda.dll' (or one of its dependencies). Try using the full path with constructor syntax.,2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/87,请检查一下cuda或cudnn是否存在安装问题,17 +18,加载txt文件失败?,2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/89,![JppHrGOWFa](https://user-images.githubusercontent.com/109277248/232009383-bf7c46d1-a01e-4e0a-9de6-5b5ed3e36158.jpg),18 +19,NameError: name 'chatglm' is not defined,2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/90,"This share link expires in 72 hours. For free permanent hosting and GPU upgrades (NEW!), check out Spaces: https://huggingface.co/spaces",19 +20,打不开地址?,2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/91,报错数据如下:,20 +21,加载md文件出错,2023-04-14.00,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/98,运行 webui.py后能访问页面,上传一个md文件后,日志中有错误。等待后能加载完成,提示可以提问了,但提问没反应,日志中有错误。 具体日志如下。,21 +22,建议增加获取在线知识的能力,2023-04-15.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/101,建议增加获取在线知识的能力,22 +23,txt 未能成功加载,2023-04-15.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/103,hinese. Creating a new one with MEAN pooling.,23 +24,pdf加载失败,2023-04-15.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/105,e:\a.txt加载成功了,e:\a.pdf加载就失败,pdf文件里面前面几页是图片,后面都是文字,加载失败没有报更多错误,请问该怎么排查?,24 +25,一直停在文本加载处,2023-04-15.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/108,一直停在文本加载处,25 +26," File ""/root/.cache/huggingface/modules/transformers_modules/chatglm-6b/modeling_chatglm.py"", line 440, in forward new_tensor_shape = mixed_raw_layer.size()[:-1] + ( TypeError: torch.Size() takes an iterable of 'int' (item 2 is 'float')",2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/113,按照最新的代码,发现,26 +27,后续会提供前后端分离的功能吗?,2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/114,类似这种https://github.com/lm-sys/FastChat/tree/main/fastchat/serve,27 +28,安装依赖报错,2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/115,(test) C:\Users\linh\Desktop\langchain-ChatGLM-master>pip install -r requirements.txt,28 +29,问特定问题会出现爆显存,2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/117,正常提问没问题。,29 +30,Expecting value: line 1 column 1 (char 0),2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/118,运行后 第一步加载配置一直报错:,30 +31,embedding https://huggingface.co/GanymedeNil/text2vec-large-chinese/tree/main是免费的,效果比对openai的如何?,2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/119,-------------------------------------------------------------------------------,31 +32,这是什么错误,在Colab上运行的。,2023-04-17.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/120,libcuda.so.1: cannot open shared object file: No such file or directory,32 +33,只想用自己的lora微调后的模型进行对话,不想加载任何本地文档,该如何调整?,2023-04-18.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/121,能出一个单独的教程吗,33 +34,"租的gpu,Running on local URL: http://0.0.0.0:7860 To create a public link, set `share=True` in `launch()`. 浏览器上访问不了???",2023-04-18.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/122,(chatglm20230401) root@autodl-container-e82d11963c-10ece0d7:~/autodl-tmp/chatglm/langchain-ChatGLM-20230418# python3.9 webui.py,34 +35,本地部署中的报错请教,2023-04-18.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/124,"您好,在本地运行langchain-ChatGLM过程中,环境及依赖的包都已经满足条件,但是运行webui.py,报错如下(运行cli_demo.py报错类似),请问是哪里出了错呢?盼望您的回复,谢谢!",35 +36,报错。The dtype of attention mask (torch.int64) is not bool,2023-04-18.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/131,The dtype of attention mask (torch.int64) is not bool,36 +37,[求助] pip install -r requirements.txt 的时候出现以下报错。。。有大佬帮忙看看怎么搞么,下的release里面的包,2023-04-18.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/134,$ pip install -r requirements.txt,37 +38,如何提升根据问题搜索到对应知识的准确率,2023-04-19.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/136,外链知识库最大的问题在于问题是短文本,知识是中长文本。如何根据问题精准的搜索到对应的知识是个最大的问题。这类本地化项目不像百度,由无数的网页,基本上每个问题都可以找到对应的页面。,38 +39,是否可以增加向量召回的阈值设定,有些召回内容相关性太低,导致模型胡言乱语,2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/140,如题,39 +40,输入长度问题,2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/141,感谢作者支持ptuning微调模型。,40 +41,已有部署好的chatGLM-6b,如何通过接口接入?,2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/144,已有部署好的chatGLM-6b,如何通过接口接入,而不是重新加载一个模型;,41 +42,执行web_demo.py后,显示Killed,就退出了,是不是配置不足呢?,2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/146,![图片](https://user-images.githubusercontent.com/26102866/233256425-c7aab999-11d7-4de9-867b-23ef18d519e4.png),42 +43,执行python cli_demo1.py,2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/147,Traceback (most recent call last):,43 +44,报错:ImportError: cannot import name 'GENERATION_CONFIG_NAME' from 'transformers.utils',2023-04-20.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/149,(mychatGLM) PS D:\Users\admin3\zrh\langchain-ChatGLM> python cli_demo.py,44 +45,上传文件并加载知识库时,会不停地出现临时文件,2023-04-21.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/153,环境:ubuntu 18.04,45 +46,向知识库中添加文件后点击”上传文件并加载知识库“后Segmentation fault报错。,2023-04-23.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/161,运行服务后的提示如下:,46 +47,langchain-serve 集成,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/162,Hey 我是来自 [langchain-serve](https://github.com/jina-ai/langchain-serve) 的dev!,47 +48,大佬们,wsl的ubuntu怎么配置用cuda加速,装了运行后发现是cpu在跑,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/164,大佬们,wsl的ubuntu怎么配置用cuda加速,装了运行后发现是cpu在跑,48 +49,在github codespaces docker运行出错,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/165,docker run -d --restart=always --name chatglm -p 7860:7860 -v /www/wwwroot/code/langchain-ChatGLM:/chatGLM chatglm,49 +50,有计划接入Moss模型嘛,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/166,后续会开展测试,目前主要在优化langchain部分效果,如果有兴趣也欢迎提PR,50 +51,怎么实现 API 部署?,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/168,利用 fastapi 实现 API 部署方式,具体怎么实现,有方法说明吗?,51 +52, 'NoneType' object has no attribute 'message_types_by_name'报错,2023-04-24.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/169,_HISTOGRAMPROTO = DESCRIPTOR.message_types_by_name['HistogramProto'],52 +53,能否指定自己训练的text2vector模型?,2023-04-25.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/172,请问大佬:,53 +54,关于项目支持的模型以及quantization_bit潜在的影响的问题,2023-04-26.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/176,作者您好~,54 +55,运行python3.9 api.py WARNING: You must pass the application as an import string to enable 'reload' or 'workers'.,2023-04-26.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/179,api.py文件最下面改成这样试试:,55 +56,ValidationError: 1 validation error for HuggingFaceEmbeddings model_kwargs extra fields not permitted (type=value_error.extra),2023-04-26.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/180,ValidationError: 1 validation error for HuggingFaceEmbeddings,56 +57,如果没有检索到相关性比较高的,回答“我不知道”,2023-04-26.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/181,如果通过设计system_template,让模型在搜索到的文档都不太相关的情况下回答“我不知道”,57 +58,请问如果不能联网,6B之类的文件从本地上传需要放到哪里,2023-04-26.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/182,感谢大佬的项目,很有启发~,58 +59,知识库问答--输入新的知识库名称是中文的话,会报error,2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/184,知识库问答--输入新的知识库名称是中文的话,会报error,选择要加载的知识库那里也不显示之前添加的知识库,59 +60,现在能通过问题匹配的相似度值,来直接返回文档中的文段,而不经过模型吗?因为有些答案在文档中,模型自己回答,不能回答文档中的答案,2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/186,现在能通过问题匹配的相似度值,来直接返回文档中的文段,而不经过模型吗?因为有些答案在文档中,模型自己回答,不能回答文档中的答案。也就是说,提供向量检索回答+模型回答相结合的策略。如果相似度值高于一定数值,直接返回文档中的文本,没有高于就返回模型的回答或者不知道,60 +61,"TypeError: The type of ChatGLM.callback_manager differs from the new default value; if you wish to change the type of this field, please use a type annotation",2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/188,"Mac 运行 python3 ./webui.py 报 TypeError: The type of ChatGLM.callback_manager differs from the new default value; if you wish to change the type of this field, please use a type annotation",61 +62,Not Enough Memory,2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/190,"运行命令行程序python cli_demo.py, 已经成功加载pdf文件, 报“DefaultCPUAllocator: not enough memory: you tried to allocate 458288380900 bytes”错误,请问哪里可以配置default memory",62 +63,参与开发问题,2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/191,1.是否需要进专门的开发群,63 +64,对话框中代码片段格式需改进,2023-04-27.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/192,最好能改进下输出代码片段的格式,目前输出的格式还不友好。,64 +65,请问未来有可能支持belle吗,2023-04-28.01,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/195,如题,谢谢大佬,65 +66,TypeError: cannot unpack non-iterable NoneType object,2023-04-28.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/200,"When i tried to change the knowledge vector store through `init_knowledge_vector_store`, the error `TypeError: cannot unpack non-iterable NoneType object` came out.",66 +67,生成结果,2023-04-28.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/202,你好,想问一下langchain+chatglm-6B,找到相似匹配的prompt,是直接返回prompt对应的答案信息,还是chatglm-6B在此基础上自己优化答案?,67 +68,在win、ubuntu下都出现这个错误:attributeerror: 't5forconditionalgeneration' object has no attribute 'stream_chat',2023-04-29.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/207,在win、ubuntu。下载完模型后,没办法修改代码以执行本地模型,每次都要重新输入路径; LLM 模型、Embedding 模型支持也都在官网下的,在其他项目(wenda)下可以使用,68 +69,[FEATURE] knowledge_based_chatglm.py: renamed or missing?,2023-04-30.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/210,"Not found. Was it renamed? Or, is it missing? How can I get it?",69 +70,sudo apt-get install -y nvidia-container-toolkit-base执行报错,2023-05-01.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/211,**问题描述 / Problem Description**,70 +71,效果不佳几乎答不上来,2023-05-01.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/212,提供了50条问答的docx文件,71 +72,有没有可能新增一个基于chatglm api调用的方式构建langchain,2023-05-02.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/218,我有两台8G GPU/40G内存的服务器,一个台做成了chatglm的api ;想基于另外一台服务器部署langchain;网上好像没有类似的代码。,72 +73,电脑是intel的集成显卡; 运行时告知我找不到nvcuda.dll,模型无法运行,2023-05-02.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/219,您好,我的电脑是intel的集成显卡,不过CPU是i5-11400 @ 2.60GHz ,内存64G;,73 +74,根据langchain官方的文档和使用模式,是否可以改Faiss为Elasticsearch?会需要做哪些额外调整?求解,2023-05-03.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/221,本人新手小白,由于业务模式的原因(有一些自己的场景和优化),希望利用Elasticsearch做这个体系内部的检索机制,不知道是否可以替换,同时,还会涉及到哪些地方的改动?或者说可能会有哪些其他影响,希望作者和大佬们不吝赐教!,74 +75,请问未来有可能支持t5吗,2023-05-04.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/224,请问可能支持基於t5的模型吗?,75 +76,[BUG] 内存溢出 / torch.cuda.OutOfMemoryError:,2023-05-04.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/229,**问题描述 / Problem Description**,76 +77,报错 No module named 'chatglm_llm',2023-05-04.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/230,明明已经安装了包,却在python里吊不出来,77 +78,能出一个api部署的描述文档吗,2023-05-04.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/233,**功能描述 / Feature Description**,78 +79,使用docs/API.md 出错,2023-05-04.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/234,使用API.md文档2种方法,出错,79 +80,加载pdf文档报错?,2023-05-05.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/238,ew one with MEAN pooling.,80 +81,上传的本地知识文件后再次上传不能显示,只显示成功了一个,别的上传成功后再次刷新就没了,2023-05-05.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/239,您好,项目有很大启发,感谢~,81 +82,创建了新的虚拟环境,安装了相关包,并且自动下载了相关的模型,但是仍旧出现:OSError: Unable to load weights from pytorch checkpoint file for,2023-05-05.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/240,![78ac8e663fdc312d0e9d78da95925c4](https://user-images.githubusercontent.com/34124260/236378728-9ea4424f-0f7f-4013-9d33-820b723de321.png),82 +83,[BUG] 数据加载不进来,2023-05-05.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/243,使用的.txt格式,utf-8编码,报以下错误,83 +84,不能读取pdf,2023-05-05.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/244,请问是webui还是cli_demo,84 +85,本地txt文件有500M,加载的时候很慢,如何提高速度?,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/251,![yayRzxSYHP](https://user-images.githubusercontent.com/109277248/236592902-f5ab338d-c1e9-43dc-ae16-9df2cd3c1378.jpg),85 +86,[BUG] gradio上传知识库后刷新之后 知识库就不见了 只有重启才能看到之前的上传的知识库,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/253,gradio上传知识库后刷新之后 知识库就不见了 只有重启才能看到之前的上传的知识库,86 +87,[FEATURE] 可以支持 OpenAI 的模型嘛?比如 GPT-3、GPT-3.5、GPT-4;embedding 增加 text-embedding-ada-002,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/254,**功能描述 / Feature Description**,87 +88,[FEATURE] 能否增加对于milvus向量数据库的支持 / Concise description of the feature,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/256,**功能描述 / Feature Description**,88 +89,CPU和GPU上跑,除了速度有区别,准确率效果回答上有区别吗?,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/259,理论上没有区别,89 +90,m1,请问在生成回答时怎么看是否使用了mps or cpu?,2023-05-06.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/260,m1,请问在生成回答时怎么看是否使用了mps or cpu?,90 +91,知识库一刷新就没了,2023-05-07.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/263,知识库上传后刷新就没了,91 +92,本地部署报没有模型,2023-05-07.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/267,建议在下载llm和embedding模型至本地后在configs/model_config中写入模型本地存储路径后再运行,92 +93,[BUG] python3: can't open file 'webui.py': [Errno 2] No such file or directory,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/269,**问题描述 / Problem Description**,93 +94,模块缺失提示,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/271,因为已有自己使用的docker环境,直接启动webui.py,提示,94 +95,"运行api.py后,执行curl -X POST ""http://127.0.0.1:7861"" 报错?",2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/272,"执行curl -X POST ""http://127.0.0.1:7861"" \ -H 'Content-Type: application/json' \ -d '{""prompt"": ""你好"", ""history"": []}',报错怎么解决",95 +96,[BUG] colab安装requirements提示protobuf版本问题?,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/273,pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.,96 +97,请问项目里面向量相似度使用了什么方法计算呀?,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/275,基本按照langchain里的FAISS.similarity_search_with_score_by_vector实现,97 +98,[BUG] 安装detectron2后,pdf无法加载,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/276,**问题描述 / Problem Description**,98 +99,[BUG] 使用ChatYuan-V2模型无法流式输出,会报错,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/277,一方面好像是ChatYuan本身不支持stream_chat,有人在clueai那边提了issue他们说还没开发,所以估计这个attribute调不起来;但是另一方面看报错好像是T5模型本身就不是decoder-only模型,所以不能流式输出吧(个人理解),99 +100,[BUG] 无法加载text2vec模型,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/278,**问题描述 / Problem Description**,100 +101,请问能否增加网络搜索功能,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/281,请问能否增加网络搜索功能,101 +102,[FEATURE] 结构化数据sql、excel、csv啥时会支持呐。,2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/283,**功能描述 / Feature Description**,102 +103,TypeError: ChatGLM._call() got an unexpected keyword argument 'stop',2023-05-08.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/284,No sentence-transformers model found with name D:\DevProject\langchain-ChatGLM\GanymedeNil\text2vec-large-chinese. Creating a new one with MEAN pooling.,103 +104,关于api.py的一些bug和设计逻辑问题?,2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/285,首先冒昧的问一下,这个api.py,开发者大佬们是在自己电脑上测试后确实没问题吗?,104 +105,有没有租用的算力平台上,运行api.py后,浏览器http://localhost:7861/报错,2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/287,是不是租用的gpu平台上都会出现这个问题???,105 +106,请问一下项目中有用到文档段落切割方法吗?,2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/288,text_load中的文档切割方法用上了吗?在代码中看好像没有用到?,106 +107,"报错 raise ValueError(f""Knowledge base {knowledge_base_id} not found"") ValueError: Knowledge base ./vector_store not found",2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/289,"File ""/root/autodl-tmp/chatglm/langchain-ChatGLM-master/api.py"", line 183, in chat",107 +108,能接入vicuna模型吗,2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/290,目前本地已经有了vicuna模型能直接接入吗?,108 +109,[BUG] 提问公式相关问题大概率爆显存,2023-05-09.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/291,**问题描述 / Problem Description**,109 +110,安装pycocotools失败,找了好多方法都不能解决。,2023-05-10.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/292,**问题描述 / Problem Description**,110 +111,使用requirements安装,PyTorch安装的是CPU版本,2023-05-10.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/294,如题目,使用requirements安装,PyTorch安装的是CPU版本,运行程序的时候,也是使用CPU在工作。,111 +112,能不能给一个毛坯服务器的部署教程,2023-05-10.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/298,“开发部署”你当成服务器的部署教程用就行了。,112 +113, Error(s) in loading state_dict for ChatGLMForConditionalGeneration:,2023-05-10.02,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/299,运行中出现的问题,7860的端口页面显示不出来,求助。,113 +114,ChatYuan-large-v2模型加载失败,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/300,**实际结果 / Actual Result**,114 +115,新增摘要功能,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/303,你好,后续会考虑新增对长文本信息进行推理和语音理解功能吗?比如生成摘要,115 +116,[BUG] pip install -r requirements.txt 出错,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/304,pip install langchain -i https://pypi.org/simple,116 +117,[BUG] 上传知识库文件报错,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/305,![19621e29eaa547d01213bee53d81e6a](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/84606552/7f6ceb46-e494-4b0e-939c-23b585a6d9d8),117 +118,[BUG] AssertionError: Component with id 41 not a valid input component.,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/306,**问题描述 / Problem Description**,118 +119,[BUG] CUDA out of memory with container deployment,2023-05-10.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/310,**问题描述 / Problem Description**,119 +120,[FEATURE] 增加微调训练功能,2023-05-11.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/311,**功能描述 / Feature Description**,120 +121,如何使用多卡部署,多个gpu,2023-05-11.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/315,"机器上有多个gpu,如何全使用了",121 +122,请问这个知识库问答,和chatglm的关系是什么,2023-05-11.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/319,这个知识库问答,哪部分关联到了chatglm,是不是没有这个chatglm,知识库问答也可单单拎出来,122 +123,[BUG] 运行的时候报错ImportError: libcudnn.so.8: cannot open shared object file: No such file or directory,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/324,**问题描述 / Problem Description**raceback (most recent call last):,123 +124,webui启动成功,但有报错,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/325,**问题描述 / Problem Description**,124 +125,切换MOSS的时候报错,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/327,danshi但是发布的源码中,,125 +126,vicuna模型是否能接入?,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/328,您好!关于MOSS模型和vicuna模型,都是AutoModelForCausalLM来加载模型的,但是稍作更改(模型路径这些)会报这个错误。这个错误的造成是什么,126 +127,你好,请问一下在阿里云CPU服务器上跑可以吗?可以的话比较理想的cpu配置是什么?,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/330,你好,请问一下在阿里云CPU服务器上跑可以吗?可以的话比较理想的cpu配置是什么?,127 +128,你好,请问8核32g的CPU可以跑多轮对话吗?,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/331,什么样的cpu配置比较好呢?我目前想部署CPU下的多轮对话?,128 +129,[BUG] 聊天内容输入超过10000个字符系统出现错误,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/332,聊天内容输入超过10000个字符系统出现错误,如下图所示:,129 +130,能增加API的多用户访问接口部署吗?,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/333,默认部署程序仅支持单用户访问,多用户则需要排队访问。测试过相关的几个Github多用户工程,但是其中一些仍然不满足要求。本节将系统介绍如何实现多用户同时访问ChatGLM的部署接口,包括http、websocket(流式输出,stream)和web页面等方式,主要目录如下所示。,130 +131,多卡部署,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/334,用单机多卡或多机多卡,fastapi部署模型,怎样提高并发,131 +132,WEBUI能否指定知识库目录?,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/335,**功能描述 / Feature Description**,132 +133,[BUG] Cannot read properties of undefined (reading 'error'),2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/336,**问题描述 / Problem Description**,133 +134,[BUG] 1 validation error for HuggingFaceEmbeddings model_kwargs extra fields not permitted.,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/337,模型加载到 100% 后出现问题:,134 +135,上传知识库需要重启能不能修复一下,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/338,挺严重的这个问题,135 +136,[BUG] 4块v100卡爆显存,在LLM会话模式也一样,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/339,**问题描述 / Problem Description**,136 +137,针对上传的文件配置不同的TextSpliter,2023-05-12.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/341,1. 目前的ChineseTextSpliter切分对英文尤其是代码文件不友好,而且限制固定长度;导致对话结果不如人意,137 +138,[FEATURE] 未来可增加Bloom系列模型吗?根据甲骨易的测试,这系列中文评测效果不错,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/346,**功能描述 / Feature Description**,138 +139,[BUG] v0.1.12打包镜像后启动webui.py失败 / Concise description of the issue,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/347,**问题描述 / Problem Description**,139 +140,切换MOSS模型时报错,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/349,昨天问了下,说是transformers版本不对,需要4.30.0,发现没有这个版本,今天更新到4.29.1,依旧报错,错误如下,140 +141,[BUG] pdf文档加载失败,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/350,**问题描述 / Problem Description**,141 +142,建议可以在后期增强一波注释,这样也有助于更多人跟进提PR,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/351,知道作者和团队在疯狂更新审查代码,只是建议后续稳定后可以把核心代码进行一些注释的补充,从而能帮助更多人了解各个模块作者的思路从而提出更好的优化。,142 +143,[FEATURE] MOSS 量化版支援,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/353,**功能描述 / Feature Description**,143 +144,[BUG] moss模型无法加载,2023-05-13.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/356,**问题描述 / Problem Description**,144 +145,[BUG] load_doc_qa.py 中的 load_file 函数有bug,2023-05-14.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/358,原函数为:,145 +146,[FEATURE] API模式,知识库加载优化,2023-05-14.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/359,如题,当前版本,每次调用本地知识库接口,都将加载一次知识库,是否有更好的方式?,146 +147,运行Python api.py脚本后端部署后,怎么使用curl命令调用?,2023-05-15.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/361,也就是说,我现在想做个对话机器人,想和公司的前后端联调?怎么与前后端相互调用呢?可私信,有偿解答!!!,147 +148,上传知识库需要重启能不能修复一下,2023-05-15.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/363,上传知识库需要重启能不能修复一下,148 +149,[BUG] pip install -r requirements.txt -i https://pypi.tuna.tsinghua.edu.cn/simple,2023-05-15.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/364,我的python是3.8.5的,149 +150,pip install gradio 报错,2023-05-15.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/367,大佬帮我一下,150 +151,[BUG] pip install gradio 一直卡不动,2023-05-15.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/369,![aba82742dd9d4d242181662eb5027a7](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/84606552/cd9600d9-f6e7-46b7-b1be-30ed8b99f76b),151 +152,[BUG] 简洁阐述问题 / Concise description of the issue,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/370,初次加载本地知识库成功,但提问后,就无法重写加载本地知识库,152 +153,[FEATURE] 简洁阐述功能 / Concise description of the feature,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/371,**功能描述 / Feature Description**,153 +154,在windows上,模型文件默认会安装到哪,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/372,-------------------------------------------------------------------------------,154 +155,[FEATURE] 兼顾对话管理,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/374,如何在知识库检索的情况下,兼顾对话管理?,155 +156,llm device: cpu embedding device: cpu,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/376,**问题描述 / Problem Description**,156 +157,[FEATURE] 简洁阐述功能 /文本文件的知识点之间使用什么分隔符可以分割?,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/377,**功能描述 / Feature Description**,157 +158,[BUG] 上传文件失败:PermissionError: [WinError 32] 另一个程序正在使用此文件,进程无法访问。,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/379,**问题描述 / Problem Description**,158 +159,[BUG] 执行python api.py 报错,2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/383,错误信息,159 +160,model_kwargs extra fields not permitted (type=value_error.extra),2023-05-16.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/384,"大家好,请问这个有遇到的么,?",160 +161,[BUG] 简洁阐述问题 / Concise description of the issue,2023-05-17.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/385,执行的时候出现了ls1 = [ls[0]],161 +162,[FEATURE] 性能优化,2023-05-17.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/388,**功能描述 / Feature Description**,162 +163,"[BUG] Moss模型问答,RuntimeError: probability tensor contains either inf, nan or element < 0",2023-05-17.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/390,**问题描述 / Problem Description**,163 +164,有没有人知道v100GPU的32G显存,会报错吗?支持V100GPU吗?,2023-05-17.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/392,**问题描述 / Problem Description**,164 +165,针对于编码问题比如'gbk' codec can't encode character '\xab' in position 14: illegal multibyte sequence粗浅的解决方法,2023-05-17.03,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/397,**功能描述 / Feature Description**,165 +166,Could not import sentence_transformers python package. Please install it with `pip install sentence_transformers`.,2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/400,**问题描述 / Problem Description**,166 +167,支持模型问答与检索问答,2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/401,不同的query,根据意图不一致,回答也应该不一样。,167 +168,文本分割的时候,能不能按照txt文件的每行进行分割,也就是按照换行符号\n进行分割???,2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/403,下面的代码应该怎么修改?,168 +169,local_doc_qa/local_doc_chat 接口响应是串行,2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/405,**问题描述 / Problem Description**,169 +170,"为什么找到出处了,但是还是无法回答该问题?",2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/406,![图片](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/3349611/1fc81d61-2409-4330-9065-fdda1a27c86a),170 +171,"请问下:知识库测试中的:添加单条内容,如果换成文本导入是是怎样的格式?我发现添加单条内容测试效果很好.",2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/412,"我发现在知识库测试中`添加单条内容`,并且勾选`禁止内容分句入库`,即使 `不开启上下文关联`的测试效果都非常好.",171 +172,[BUG] 无法配置知识库,2023-05-18.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/413,**问题描述 / Problem Description**,172 +173,[BUG] 部署在阿里PAI平台的EAS上访问页面是白屏,2023-05-19.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/414,**问题描述 / Problem Description**,173 +174,API部署后调用/local_doc_qa/local_doc_chat 返回Knowledge base samples not found,2023-05-19.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/416,入参,174 +175,[FEATURE] 上传word另存为的txt文件报 'ascii' codec can't decode byte 0xb9 in position 6: ordinal not in range(128),2023-05-20.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/421,上传word另存为的txt文件报,175 +176,创建保存的知识库刷新后没有出来,这个知识库是永久保存的吗?可以连外部的 向量知识库吗?,2023-05-21.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/422,创建保存的知识库刷新后没有出来,这个知识库是永久保存的吗?可以连外部的 向量知识库吗?,176 +177,[BUG] 用colab运行,无法加载模型,报错:'NoneType' object has no attribute 'message_types_by_name',2023-05-21.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/423,**问题描述 / Problem Description**,177 +178,请问是否需要用到向量数据库?以及什么时候需要用到向量数据库?,2023-05-21.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/424,目前用的是 text2vec , 请问是否需要用到向量数据库?以及什么时候需要用到向量数据库?,178 +179,huggingface模型引用问题,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/427,它最近似乎变成了一个Error?,179 +180,你好,加载本地txt文件出现这个killed错误,TXT文件有100M左右大小。原因是?谢谢。,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/429,"",180 +181,想请问一下,关于对本地知识的管理是如何管理?例如:通过http API接口添加数据 或者 删除某条数据,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/430,例如:通过http API接口添加、删除、修改 某条数据。,181 +182,[FEATURE] 双栏pdf识别问题,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/432,试了一下模型,感觉对单栏pdf识别的准确性较高,但是由于使用的基本是ocr的技术,对一些双栏pdf论文识别出来有很多问题,请问有什么办法改善吗?,182 +183,部署启动小问题,小弟初学求大佬解答,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/433,1.python loader/image_loader.py时,提示ModuleNotFoundError: No module named 'configs',但是跑python webui.py还是还能跑,183 +184,能否支持检测到目录下文档有增加而去增量加载文档,不影响前台对话,其实就是支持读写分离。如果能支持查询哪些文档向量化了,删除过时文档等就更好了,谢谢。,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/434,**功能描述 / Feature Description**,184 +185,[BUG] 简洁阐述问题 / windows 下cuda错误,请用https://github.com/Keith-Hon/bitsandbytes-windows.git,2023-05-22.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/435,pip install git+https://github.com/Keith-Hon/bitsandbytes-windows.git,185 +186,"[BUG] from commit 33bbb47, Required library version not found: libbitsandbytes_cuda121_nocublaslt.so. Maybe you need to compile it from source?",2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/438,**问题描述 / Problem Description**,186 +187,[BUG] 简洁阐述问题 / Concise description of the issue上传60m的txt文件报错,显示超时,请问这个能上传的文件大小有限制吗,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/439,"ERROR 2023-05-23 11:13:09,627-1d: Timeout reached while detecting encoding for ./docs/GLM模型格式数据.txt",187 +188,[BUG] TypeError: issubclass() arg 1 must be a class,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/440,**问题描述**,188 +189,"执行python3 webui.py后,一直提示”模型未成功加载,请到页面左上角""模型配置""选项卡中重新选择后点击""加载模型""按钮“",2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/441,**问题描述 / Problem Description**,189 +190,是否能提供网页文档得导入支持,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/444,现在很多都是在线文档作为协作得工具,所以通过URL导入在线文档需求更大,190 +191,[BUG] history 索引问题,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/445,在比较对话框的history和模型chat function 中的history时, 发现并不匹配,在传入 llm._call 时,history用的索引是不是有点问题,导致上一轮对话的内容并不输入给模型。,191 +192,[BUG] moss_llm没有实现,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/447,有些方法没支持,如history_len,192 +193,请问langchain-ChatGLM如何删除一条本地知识库的数据?,2023-05-23.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/448,例如:用户刚刚提交了一条错误的数据到本地知识库中了,现在如何在本地知识库从找到,并且对此删除。,193 +194,[BUG] 简洁阐述问题 / UnboundLocalError: local variable 'resp' referenced before assignment,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/450,"在最新一版的代码中, 运行api.py 出现了以上错误(UnboundLocalError: local variable 'resp' referenced before assignment), 通过debug的方式观察到local_doc_qa.llm.generatorAnswer(prompt=question, history=history,streaming=True)可能不返回任何值。",194 +195,请问有没有 PROMPT_TEMPLATE 能让模型不回答敏感问题,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/452,## PROMPT_TEMPLATE问题,195 +196,[BUG] 测试环境 Python 版本有误,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/456,**问题描述 / Problem Description**,196 +197,[BUG] webui 部署后样式不正确,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/458,**问题描述 / Problem Description**,197 +198,配置默认LLM模型的问题,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/459,**问题描述 / Problem Description**,198 +199,[FEATURE]是时候更新一下autoDL的镜像了,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/460,如题,跑了下autoDL的镜像,发现是4.27号的,git pull新版本的代码功能+老的依赖环境,各种奇奇怪怪的问题。,199 +200,[BUG] tag:0.1.13 以cpu模式下,想使用本地模型无法跑起来,各种路径参数问题,2023-05-24.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/462,-------------------------------------------------------------------------------,200 +201,[BUG] 有没有同学遇到过这个错!!!加载本地txt文件出现这个killed错误,TXT文件有100M左右大小。,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/463,运行cli_demo.py。是本地的txt文件太大了吗?100M左右。,201 +202,API版本能否提供WEBSOCKET的流式接口,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/464,webui 版本中,采用了WS的流式输出,整体感知反应很快,202 +203,[BUG] 安装bug记录,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/465,按照[install文档](https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/docs/INSTALL.md)安装的,,203 +204,VUE的pnmp i执行失败的修复-用npm i命令即可,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/466,感谢作者!非常棒的应用,用的很开心。,204 +205,请教个问题,有没有人知道cuda11.4是否支持???,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/467,请教个问题,有没有人知道cuda11.4是否支持???,205 +206,请问有实现多轮问答中基于问题的搜索上下文关联么,2023-05-25.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/468,在基于知识库的多轮问答中,第一个问题讲述了一个主题,后续的问题描述没有包含这个主题的关键词,但又存在上下文的关联。如果用后续问题去搜索知识库有可能会搜索出无关的信息,从而导致大模型无法正确回答问题。请问这个项目要考虑这种情况吗?,206 +207,[BUG] 内存不足的问题,2023-05-26.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/470,我用了本地的chatglm-6b-int4模型,然后显示了内存不足(win10+32G内存+1080ti11G),一般需要多少内存才足够?这个bug应该如何解决?,207 +208,[BUG] 纯内网环境安装pycocotools失败,2023-05-26.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/472,**问题描述 / Problem Description**,208 +209,[BUG] webui.py 重新加载模型会导致 KeyError,2023-05-26.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/473,**问题描述 / Problem Description**,209 +210,chatyuan无法使用,2023-05-26.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/475,**问题描述 / Problem Description**,210 +211,[BUG] 文本分割模型AliTextSplitter存在bug,会把“.”作为分割符,2023-05-26.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/476,"阿里达摩院的语义分割模型存在bug,默认会把"".”作为分割符进行分割而不管上下文语义。是否还有其他分割符则未知。建议的修改方案:把“.”统一替换为其他字符,分割后再替换回来。或者添加其他分割模型。",211 +212,[BUG] RuntimeError: Error in faiss::FileIOReader::FileIOReader(const char*) a,2023-05-27.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/479,**问题描述 / Problem Description**,212 +213,[FEATURE] 安装,为什么conda create要额外指定路径 用-p ,而不是默认的/envs下面,2023-05-28.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/481,##**功能描述 / Feature Description**,213 +214,[小白求助] 通过Anaconda执行webui.py后,无法打开web链接,2023-05-28.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/485,在执行webui.py命令后,http://0.0.0.0:7860复制到浏览器后无法打开,显示“无法访问此网站”。,214 +215,[BUG] 使用 p-tuningv2后的模型,重新加载报错,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/486,把p-tunningv2训练完后的相关文件放到了p-tunningv2文件夹下,勾选使用p-tuningv2点重新加载模型,控制台输错错误信息:,215 +216,[小白求助] 服务器上执行webui.py后,在本地无法打开web链接,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/487,此项目执行在xxx.xx.xxx.xxx服务器上,我在webui.py上的代码为 (demo,216 +217,[FEATURE] 能不能支持VisualGLM-6B,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/488,**功能描述 / Feature Description**,217 +218,你好,问一下各位,后端api部署的时候,支持多用户同时问答吗???,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/489,支持多用户的话,最多支持多少用户问答?根据硬件而定吧?,218 +219,V100GPU显存占满,而利用率却为0,这是为什么?,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/491,"",219 +220,[求助] 如果在公司内部搭建产品知识库,使用INT-4模型,200人规模需要配置多少显存的服务器?,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/492,如题,计划给公司搭一个在线知识库。,220 +221,你好,请教个问题,目前问答回复需要20秒左右,如何提高速度?V10032G服务器。,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/493,**问题描述 / Problem Description**,221 +222,[FEATURE] 如何实现只匹配下文,而不要上文的结果,2023-05-29.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/494,在构建自己的知识库时,主要采用问答对的形式,那么也就是我需要的回答是在我的问题下面的内容,但是目前设置了chunk_size的值以后匹配的是上下文的内容,但我实际并不需要上文的。为了实现更完整的展示下面的答案,我只能调大chunk_size的值,但实际上上文的一半内容都是我不需要的。也就是扔了一半没用的东西给prompt,在faiss.py中我也没找到这块的一些描述,请问该如何进行修改呢?,222 +223,你好,问一下,我调用api.py部署,为什么用ip加端口可以使用postman调用,而改为域名使用postman无法调用?,2023-05-30.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/497,![5ufBSWxLyF](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/109277248/70e2fbac-5699-48d0-b0d1-3dc84fd042c2),223 +224,调用api.py中的stream_chat,返回source_documents中出现中文乱码。,2023-05-30.04,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/498,-------------------------------------------------------------------------------,224 +225,[BUG] 捉个虫,api.py中的stream_chat解析json问题,2023-05-30.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/501,**问题描述 / Problem Description**,225 +226,windows本地部署遇到了omp错误,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/502,**问题描述 / Problem Description**,226 +227,"[BUG] bug14 ,""POST /local_doc_qa/upload_file HTTP/1.1"" 422 Unprocessable Entity",2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/503,上传的文件报错,返回错误,api.py,227 +228,你好,请教个问题,api.py部署的时候,如何改为多线程调用?谢谢,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/505,目前的api.py脚本不支持多线程,228 +229,你好,请教一下。api.py部署的时候,能不能提供给后端流失返回结果。,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/507,curl -X 'POST' \,229 +230,流式输出,流式接口,使用server-sent events技术。,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/508,想这样一样,https://blog.csdn.net/weixin_43228814/article/details/130063010,230 +231,计划增加流式输出功能吗?ChatGLM模型通过api方式调用响应时间慢怎么破,Fastapi流式接口来解惑,能快速提升响应速度,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/509,**问题描述 / Problem Description**,231 +232,[BUG] 知识库上传时发生ERROR (could not open xxx for reading: No such file or directory),2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/510,**问题描述 / Problem Description**,232 +233,api.py脚本打算增加SSE流式输出吗?,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/511,curl调用的时候可以检测第一个字,从而提升回复的体验,233 +234,[BUG] 使用tornado实现webSocket,可以多个客户端同时连接,并且实现流式回复,但是多个客户端同时使用,答案就很乱,是模型不支持多线程吗,2023-05-31.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/512,import asyncio,234 +235,支持 chinese_alpaca_plus_lora 吗 基于llama的,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/514,支持 chinese_alpaca_plus_lora 吗 基于llama的,https://github.com/ymcui/Chinese-LLaMA-Alpaca这个项目的,235 +236,[BUG] 现在能读图片的pdf了,但是文字的pdf反而读不了了,什么情况???,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/515,**问题描述 / Problem Description**,236 +237,在推理的过程中卡住不动,进程无法正常结束,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/516,**问题描述 / Problem Description**,237 +238,curl调用的时候,从第二轮开始,curl如何传参可以实现多轮对话?,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/517,第一轮调用:,238 +239,建议添加api.py部署后的日志管理功能?,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/518,-------------------------------------------------------------------------------,239 +240,有大佬知道,怎么多线程部署api.py脚本吗?,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/519,api.py部署后,使用下面的请求,时间较慢,好像是单线程,如何改为多线程部署api.py:,240 +241,[BUG] 上传文件到知识库 任何格式与内容都永远失败,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/520,上传知识库的时候,传txt无法解析,就算是穿content/sample里的样例txt也无法解析,上传md、pdf等都无法加载,会持续性等待,等到了超过30分钟也不行。,241 +242,关于prompt_template的问题,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/521,请问这段prompt_template是什么意思,要怎么使用?可以给一个具体模板参考下吗?,242 +243,[BUG] 简洁阐述问题 / Concise description of the issue,2023-06-01.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/522,**问题描述 / Problem Description**,243 +244,"中文分词句号处理(关于表达金额之间的""."")",2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/523,建议处理12.6亿元的这样的分词,最好别分成12 和6亿这样的,需要放到一起,244 +245,ImportError: cannot import name 'inference' from 'paddle',2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/526,在网上找了一圈,有说升级paddle的,我做了还是没有用,有说安装paddlepaddle的,我找了豆瓣的镜像源,但安装报错cannot detect archive format,245 +246,[BUG] webscoket 接口串行问题(/local_doc_qa/stream-chat/{knowledge_base_id}),2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/527,**问题描述 / Problem Description**,246 +247,[FEATURE] 刷新页面更新知识库列表,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/528,**功能描述以及改进方案**,247 +248,[BUG] 使用ptuning微调模型后,问答效果并不好,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/530,### 未调用ptuning,248 +249,[BUG] 多轮对话效果不佳,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/532,在进行多轮对话的时候,无论设置的history_len是多少,效果都不好。事实上我将其设置成了最大值10,但在对话中,仍然无法实现多轮对话:,249 +250,"RuntimeError: MPS backend out of memory (MPS allocated: 18.00 GB, other allocations: 4.87 MB, max allowed: 18.13 GB)",2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/533,**问题描述**,250 +251, 请大家重视这个issue!真正使用肯定是多用户并发问答,希望增加此功能!!!,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/534,这得看你有多少显卡,251 +252,在启动项目的时候如何使用到多张gpu啊?,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/535,**在启动项目的时候如何使用到多张gpu啊?**,252 +253, 使用流式输出的时候,curl调用的格式是什么?,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/536,"app.websocket(""/local_doc_qa/stream-chat/{knowledge_base_id}"")(stream_chat)中的knowledge_base_id应该填什么???",253 +254,使用本地 vicuna-7b模型启动错误,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/538,环境: ubuntu 22.04 cuda 12.1 没有安装nccl,使用rtx2080与m60显卡并行计算,254 +255,为什么会不调用GPU直接调用CPU呢,2023-06-02.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/539,我的阿里云配置是16G显存,用默认代码跑webui.py时提示,255 +256,上传多个文件时会互相覆盖,2023-06-03.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/541,1、在同一个知识库中上传多个文件时会互相覆盖,无法结合多个文档的知识,有大佬知道怎么解决吗?,256 +257,[BUG] ‘gcc’不是内部或外部命令/LLM对话只能持续一轮,2023-06-03.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/542,No compiled kernel found.,257 +258,以API模式启动项目却没有知识库的接口列表?,2023-06-04.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/544,请问如何获取知识库的接口列表?如果没有需要自行编写的话,可不可以提供相关的获取方式,感谢,258 +259,程序以API模式启动的时候,如何才能让接口以stream模式被调用呢?,2023-06-05.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/546,作者您好,我在以API模式进行程序启动后,我发现接口响应时间很长,怎么样才能让接口以stream模式被调用呢?我想实现像webui模式的回答那样,259 +260,关于原文中表格转为文本后数据相关度问题。,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/547,原文中表格数据转换为文本,以 (X-Y:值;...) 的格式每一行组织成一句话,但这样做后发现相关度较低,效果很差,有何好的方案吗?,260 +261,启动后LLM和知识库问答模式均只有最后一轮记录,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/548,拉取最新代码,问答时,每次页面只显示最后一次问答记录,需要修改什么参数才可以保留历史记录?,261 +262,提供system message配置,以便于让回答不要超出知识库范围,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/549,**功能描述 / Feature Description**,262 +263,[BUG] 使用p-tunningv2报错,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/551,按照readme的指示把p-tunningv2训练完后的文件放到了p-tunningv2文件夹下,勾选使用p-tuningv2点重新加载模型,控制台提示错误信息:,263 +264,[BUG] 智障,这么多问题,也好意思放出来,浪费时间,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/553,。。。,264 +265,[FEATURE] 我看代码文件中有一个ali_text_splitter.py,为什么不用他这个文本分割器了?,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/554,我看代码文件中有一个ali_text_splitter.py,为什么不用他这个文本分割器了?,265 +266,加载文档函数报错,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/557,"def load_file(filepath, sentence_size=SENTENCE_SIZE):",266 +267,参考指引安装docker后,运行cli_demo.py,提示killed,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/558,root@b3d1bd08095c:/chatGLM# python3 cli_demo.py,267 +268,注意:如果安装错误,注意这两个包的版本 wandb==0.11.0 protobuf==3.18.3,2023-06-06.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/559,Error1: 如果启动异常报错 `protobuf` 需要更新到 `protobuf==3.18.3 `,268 +269,知识库对长文的知识相关度匹配不太理想有何优化方向,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/563,我们可能录入一个文章有 1W 字,里面涉及这个文章主题的很多角度问题,我们针对他提问,他相关度匹配的内容和实际我们需要的答案相差很大怎么办。,269 +270,使用stream-chat函数进行流式输出的时候,能使用curl调用吗?,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/565,为什么下面这样调用会报错???,270 +271,有大佬实践过 并行 或者 多线程 的部署方案吗?,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/566,+1,271 +272,多线程部署遇到问题?,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/567,"",272 +273,[BUG] 用fastchat加载vicuna-13b模型进行知识库的问答有token的限制错误,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/569,当我开启fastchat的vicuna-13b的api服务,然后config那里配置好(api本地测试过可以返回结果),然后知识库加载好之后(知识库大概有1000多个文档,用chatGLM可以正常推理),进行问答时出现token超过限制,就问了一句hello;,273 +274,现在的添加知识库,文件多了总是报错,也不知道自己加载了哪些文件,报错后也不知道是全部失败还是一部分成功;希望能有个加载指定文件夹作为知识库的功能,2023-06-07.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/574,**功能描述 / Feature Description**,274 +275,[BUG] moss模型本地加载报错,2023-06-08.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/577,moss模型本地加载报错:,275 +276,加载本地moss模型报错Can't instantiate abstract class MOSSLLM with abstract methods _history_len,2023-06-08.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/578,(vicuna) ps@ps[13:56:20]:/data/chat/langchain-ChatGLM2/langchain-ChatGLM-0.1.13$ python webui.py --model-dir local_models --model moss --no-remote-model,276 +277,[FEATURE] 能增加在前端页面控制prompt_template吗?或是能支持前端页面选择使用哪个prompt?,2023-06-08.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/579,目前只能在config里修改一个prompt,想在多个不同场景切换比较麻烦,277 +278,[BUG] streamlit ui的bug,在增加知识库时会报错,2023-06-08.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/580,**问题描述 / Problem Description**,278 +279,[FEATURE] webui/webui_st可以支持history吗?目前仅能一次对话,2023-06-08.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/581,试了下webui和webui_st都不支持历史对话啊,只能对话一次,不能默认开启所有history吗?,279 +280,启动python cli_demo.py --model chatglm-6b-int4-qe报错,2023-06-09.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/585,下载好模型,和相关依赖环境,之间运行`python cli_demo.py --model chatglm-6b-int4-qe`报错了:,280 +281,重新构建知识库报错,2023-06-09.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/586,**问题描述 / Problem Description**,281 +282,[FEATURE] 能否屏蔽paddle,我不需要OCR,效果差依赖环境还很复杂,2023-06-09.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/587,希望能不依赖paddle,282 +283,question :文档向量化这个可以自己手动实现么?,2023-06-09.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/589,现有公司级数据500G+,需要使用这个功能,请问如何手动实现这个向量化,然后并加载,283 +284,view前端能进行流式的返回吗??,2023-06-09.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/590,view前端能进行流式的返回吗??,284 +285,"[BUG] Load parallel cpu kernel failed, using default cpu kernel code",2023-06-11.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/594,**问题描述 / Problem Description**,285 +286,[BUG] 简洁阐述问题 / Concise description of the issue,2023-06-11.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/595,**问题描述 / Problem Description**,286 +287,我在上传本地知识库时提示KeyError: 'name'错误,本地知识库都是.txt文件,文件数量大约是2000+。,2023-06-12.05,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/597,"",287 +288,model_config.py中有vicuna-13b-hf模型的配置信息,但是好像还是不可用?,2023-06-12.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/600,@dongyihua543,288 +289,"ImportError: Using SOCKS proxy, but the 'socksio' package is not installed. Make sure to install httpx using `pip install httpx[socks]`.",2023-06-12.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/605,应该代理问题,但是尝试了好多方法都解决不了,,289 +290,[BUG] similarity_search_with_score_by_vector在找不到匹配的情况下出错,2023-06-12.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/607,在设置匹配阈值 VECTOR_SEARCH_SCORE_THRESHOLD 的情况下,vectorstore会返回空,此时上述处理函数会出错,290 +291,[FEATURE] 请问如何搭建英文知识库呢,2023-06-12.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/609,**功能描述 / Feature Description**,291 +292,谁有vicuna权重?llama转换之后的,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/611,**问题描述 / Problem Description**,292 +293,[FEATURE] API能实现上传文件夹的功能么?,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/612,用户懒得全选所有的文件,就想上传个文件夹,请问下API能实现这个功能么?,293 +294,请问在多卡部署后,上传单个文件作为知识库,用的是单卡在生成向量还是多卡?,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/614,目前我检测我本地多卡部署的,好像生成知识库向量的时候用的还是单卡,294 +295,[BUG] python webui.py提示非法指令,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/615,(/data/conda-langchain [root@chatglm langchain-ChatGLM]# python webui.py,295 +296,知识库文件跨行切分问题,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/616,我的知识库文件txt文件,是一行一条知识,用\n分行。,296 +297,[FEATURE] bing搜索问答有流式的API么?,2023-06-13.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/617,web端是有这个bing搜索回答,但api接口没有发现,大佬能给个提示么?,297 +298,希望出一个macos m2的安装教程,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/620,mac m2安装,模型加载成功了,知识库文件也上传成功了,但是一问答就会报错,报错内容如下,298 +299,为【出处】提供高亮显示,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/621,具体出处里面,对相关的内容高亮显示,不包含前后文。,299 +300,[BUG] CPU运行cli_demo.py,不回答,hang住,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/622,没有GPU;32G内存的ubuntu机器。,300 +301,关于删除知识库里面的文档后,LLM知识库对话的时候还是会返回该被删除文档的内容,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/623,如题,在vue前端成功执行删除知识库里面文档A.txt后,未能也在faiss索引中也删除该文档,LLM还是会返回这个A.txt的内容,并且以A.txt为出处,未能达到删除的效果,301 +302,"[BUG] 调用知识库进行问答,显存会一直叠加",2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/625,"14G的显存,调用的chatglm-6b-int8模型,进行知识库问答时,最多问答四次就会爆显存了,观察了一下显存使用情况,每一次使用就会增加一次显存,请问这样是正常的吗?是否有什么配置需要开启可以解决这个问题?例如进行一次知识库问答清空上次问题的显存?",302 +303,[BUG] web页面 重新构建数据库 失败,导致 原来的上传的数据库都没了,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/626,web页面 重新构建数据库 失败,导致 原来的上传的数据库都没了,303 +304,在CPU上运行webui.py报错Tensor on device cpu is not on the expected device meta!,2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/627,在CPU上运行python webui.py能启动,但最后有:RuntimeError: Tensor on device cpu is not on the expected device meta!,304 +305,"OSError: [WinError 1114] 动态链接库(DLL)初始化例程失败。 Error loading ""E:\xxx\envs\langchain\lib\site-packages\torch\lib\caffe2_nvrtc.dll"" or one of its dependencies.哪位大佬知道如何解决吗?",2023-06-14.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/629,**问题描述 / Problem Description**,305 +306,[BUG] WEBUI删除知识库文档,会导致知识库问答失败,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/632,如题,从知识库已有文件中选择要删除的文件,点击删除后,在问答框输入内容回车报错,306 +307,更新后的版本中,删除知识库中的文件,再提问出现error错误,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/634,针对更新版本,识别到一个问题,过程如下:,307 +308,我配置好了环境,想要实现本地知识库的问答?可是它返回给我的,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/637,没有总结,只有相关度的回复,但是我看演示里面表现的,回复是可以实现总结的,我去查询代码,308 +309,[BUG] NPM run dev can not successfully start the VUE frontend,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/638,**问题描述 / Problem Description**,309 +310,[BUG] 简洁阐述问题 / Concise description of the issue,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/639,**问题描述 / Problem Description**,310 +311,提一个模型加载的bug,我在截图中修复了,你们有空可以看一下。,2023-06-15.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/642,![model_load_bug](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/59411575/4432adc4-ccdd-45d9-aafc-5f2d1963403b),311 +312,[求助]关于设置embedding model路径的问题,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/643,如题,我之前成功跑起来过一次,但因环境丢失重新配置 再运行webui就总是报错,312 +313,Lora微调后的模型可以直接使用吗,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/646,看model_config.py里是有USE_LORA这个参数的,但是在cli_demo.py和webui.py这两个里面都没有用到,实际测试下来模型没有微调的效果,想问问现在这个功能实现了吗,313 +314,write_check_file在tmp_files目录下生成的load_file.txt是否需要一直保留,占用空间很大,在建完索引后能否删除,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/647,**功能描述 / Feature Description**,314 +315,[BUG] /local_doc_qa/list_files?knowledge_base_id=test删除知识库bug,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/649,1.新建test知识库并上传文件(在vue前端完成并检查后端发现确实生成了test文件夹以及下面的content和vec_store,315 +316,[BUG] vue webui无法加载知识库,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/650,拉取了最新的代码,分别运行了后端api和前端web,点击知识库,始终只能显示simple,无法加载知识库,316 +317,不能本地加载moss模型吗?,2023-06-16.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/652,手动下载模型设置local_model_path路径依旧提示缺少文件,该如何正确配置?,317 +318,macos m2 pro docker 安装失败,2023-06-17.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/654,macos m2 pro docker 安装失败,318 +319, [BUG] mac m1 pro 运行提示 zsh: segmentation fault,2023-06-17.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/655,运行: python webui.py,319 +320,安装 requirements 报错,2023-06-17.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/656,(langchainchatglm) D:\github\langchain-ChatGLM>pip install -r requirements.txt -i https://pypi.tuna.tsinghua.edu.cn/simple/,320 +321,[BUG] AssertionError,2023-06-17.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/658,**问题描述 / Problem Description**,321 +322,[FEATURE] 支持AMD win10 本地部署吗?,2023-06-18.06,https://github.com/imClumsyPanda/langchain-ChatGLM/issues/660,**功能描述 / Feature Description**,322 diff --git a/knowledge_base/samples/isssues_merge/langchain-ChatGLM_open.jsonl b/knowledge_base/samples/isssues_merge/langchain-ChatGLM_open.jsonl index 7a5ced26..a34816e8 100644 --- a/knowledge_base/samples/isssues_merge/langchain-ChatGLM_open.jsonl +++ b/knowledge_base/samples/isssues_merge/langchain-ChatGLM_open.jsonl @@ -1,323 +1,323 @@ -{"title": "效果如何优化", "file": "2023-04-04.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/14", "detail": "如图所示,将该项目的README.md和该项目结合后,回答效果并不理想,请问可以从哪些方面进行优化", "id": 0} -{"title": "怎么让模型严格根据检索的数据进行回答,减少胡说八道的回答呢", "file": "2023-04-04.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/15", "detail": "举个例子:", "id": 1} -{"title": "When I try to run the `python knowledge_based_chatglm.py`, I got this error in macOS(M1 Max, OS 13.2)", "file": "2023-04-07.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/32", "detail": "```python", "id": 2} -{"title": "萌新求教大佬怎么改成AMD显卡或者CPU?", "file": "2023-04-10.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/48", "detail": "把.cuda()去掉就行", "id": 3} -{"title": "输出answer的时间很长,是否可以把文本向量化的部分提前做好存储起来?", "file": "2023-04-10.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/50", "detail": "GPU:4090 24G显存", "id": 4} -{"title": "报错Use `repo_type` argument if needed.", "file": "2023-04-11.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/57", "detail": "Traceback (most recent call last):", "id": 5} -{"title": "无法打开gradio的页面", "file": "2023-04-11.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/58", "detail": "$ python webui.py", "id": 6} -{"title": "支持word,那word里面的图片正常显示吗?", "file": "2023-04-12.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/60", "detail": "如题,刚刚从隔壁转过来的,想先了解下", "id": 7} -{"title": "detectron2 is not installed. Cannot use the hi_res partitioning strategy. Falling back to partitioning with the fast strategy.", "file": "2023-04-12.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/63", "detail": "能够正常的跑起来,在加载content文件夹中的文件时,每加载一个文件都会提示:", "id": 8} -{"title": "cpu上运行webui,step3 asking时报错", "file": "2023-04-12.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/66", "detail": "web运行,文件加载都正常,asking时报错", "id": 9} -{"title": "建议弄一个插件系统", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/67", "detail": "如题弄成stable-diffusion-webui那种能装插件,再开一个存储库给使用者或插件开发,存储或下载插件。", "id": 10} -{"title": "请教加载模型出错!?", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/75", "detail": "AttributeError: module 'transformers_modules.chatglm-6b.configuration_chatglm' has no attribute 'ChatGLMConfig 怎么解决呀", "id": 11} -{"title": "从本地知识检索内容的时候,是否可以设置相似度阈值,小于这个阈值的内容不返回,即使会小于设置的VECTOR_SEARCH_TOP_K参数呢?谢谢大佬", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/76", "detail": "比如 问一些 你好/你是谁 等一些跟本地知识库无关的问题", "id": 12} -{"title": "如何改成多卡推理?", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/77", "detail": "+1", "id": 13} -{"title": "能否弄个懒人包,可以一键体验?", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/78", "detail": "能否弄个懒人包,可以一键体验?", "id": 14} -{"title": "连续问问题会导致崩溃", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/79", "detail": "看上去不是爆内存的问题,连续问问题后,会出现如下报错", "id": 15} -{"title": "AttributeError: 'NoneType' object has no attribute 'as_retriever'", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/86", "detail": "环境:windows 11, anaconda/python 3.8", "id": 16} -{"title": "FileNotFoundError: Could not find module 'nvcuda.dll' (or one of its dependencies). Try using the full path with constructor syntax.", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/87", "detail": "请检查一下cuda或cudnn是否存在安装问题", "id": 17} -{"title": "加载txt文件失败?", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/89", "detail": "![JppHrGOWFa](https://user-images.githubusercontent.com/109277248/232009383-bf7c46d1-a01e-4e0a-9de6-5b5ed3e36158.jpg)", "id": 18} -{"title": "NameError: name 'chatglm' is not defined", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/90", "detail": "This share link expires in 72 hours. For free permanent hosting and GPU upgrades (NEW!), check out Spaces: https://huggingface.co/spaces", "id": 19} -{"title": "打不开地址?", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/91", "detail": "报错数据如下:", "id": 20} -{"title": "加载md文件出错", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/98", "detail": "运行 webui.py后能访问页面,上传一个md文件后,日志中有错误。等待后能加载完成,提示可以提问了,但提问没反应,日志中有错误。 具体日志如下。", "id": 21} -{"title": "建议增加获取在线知识的能力", "file": "2023-04-15.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/101", "detail": "建议增加获取在线知识的能力", "id": 22} -{"title": "txt 未能成功加载", "file": "2023-04-15.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/103", "detail": "hinese. Creating a new one with MEAN pooling.", "id": 23} -{"title": "pdf加载失败", "file": "2023-04-15.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/105", "detail": "e:\\a.txt加载成功了,e:\\a.pdf加载就失败,pdf文件里面前面几页是图片,后面都是文字,加载失败没有报更多错误,请问该怎么排查?", "id": 24} -{"title": "一直停在文本加载处", "file": "2023-04-15.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/108", "detail": "一直停在文本加载处", "id": 25} -{"title": " File \"/root/.cache/huggingface/modules/transformers_modules/chatglm-6b/modeling_chatglm.py\", line 440, in forward new_tensor_shape = mixed_raw_layer.size()[:-1] + ( TypeError: torch.Size() takes an iterable of 'int' (item 2 is 'float')", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/113", "detail": "按照最新的代码,发现", "id": 26} -{"title": "后续会提供前后端分离的功能吗?", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/114", "detail": "类似这种https://github.com/lm-sys/FastChat/tree/main/fastchat/serve", "id": 27} -{"title": "安装依赖报错", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/115", "detail": "(test) C:\\Users\\linh\\Desktop\\langchain-ChatGLM-master>pip install -r requirements.txt", "id": 28} -{"title": "问特定问题会出现爆显存", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/117", "detail": "正常提问没问题。", "id": 29} -{"title": "Expecting value: line 1 column 1 (char 0)", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/118", "detail": "运行后 第一步加载配置一直报错:", "id": 30} -{"title": "embedding https://huggingface.co/GanymedeNil/text2vec-large-chinese/tree/main是免费的,效果比对openai的如何?", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/119", "detail": "-------------------------------------------------------------------------------", "id": 31} -{"title": "这是什么错误,在Colab上运行的。", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/120", "detail": "libcuda.so.1: cannot open shared object file: No such file or directory", "id": 32} -{"title": "只想用自己的lora微调后的模型进行对话,不想加载任何本地文档,该如何调整?", "file": "2023-04-18.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/121", "detail": "能出一个单独的教程吗", "id": 33} -{"title": "租的gpu,Running on local URL: http://0.0.0.0:7860 To create a public link, set `share=True` in `launch()`. 浏览器上访问不了???", "file": "2023-04-18.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/122", "detail": "(chatglm20230401) root@autodl-container-e82d11963c-10ece0d7:~/autodl-tmp/chatglm/langchain-ChatGLM-20230418# python3.9 webui.py", "id": 34} -{"title": "本地部署中的报错请教", "file": "2023-04-18.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/124", "detail": "您好,在本地运行langchain-ChatGLM过程中,环境及依赖的包都已经满足条件,但是运行webui.py,报错如下(运行cli_demo.py报错类似),请问是哪里出了错呢?盼望您的回复,谢谢!", "id": 35} -{"title": "报错。The dtype of attention mask (torch.int64) is not bool", "file": "2023-04-18.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/131", "detail": "The dtype of attention mask (torch.int64) is not bool", "id": 36} -{"title": "[求助] pip install -r requirements.txt 的时候出现以下报错。。。有大佬帮忙看看怎么搞么,下的release里面的包", "file": "2023-04-18.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/134", "detail": "$ pip install -r requirements.txt", "id": 37} -{"title": "如何提升根据问题搜索到对应知识的准确率", "file": "2023-04-19.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/136", "detail": "外链知识库最大的问题在于问题是短文本,知识是中长文本。如何根据问题精准的搜索到对应的知识是个最大的问题。这类本地化项目不像百度,由无数的网页,基本上每个问题都可以找到对应的页面。", "id": 38} -{"title": "是否可以增加向量召回的阈值设定,有些召回内容相关性太低,导致模型胡言乱语", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/140", "detail": "如题", "id": 39} -{"title": "输入长度问题", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/141", "detail": "感谢作者支持ptuning微调模型。", "id": 40} -{"title": "已有部署好的chatGLM-6b,如何通过接口接入?", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/144", "detail": "已有部署好的chatGLM-6b,如何通过接口接入,而不是重新加载一个模型;", "id": 41} -{"title": "执行web_demo.py后,显示Killed,就退出了,是不是配置不足呢?", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/146", "detail": "![图片](https://user-images.githubusercontent.com/26102866/233256425-c7aab999-11d7-4de9-867b-23ef18d519e4.png)", "id": 42} -{"title": "执行python cli_demo1.py", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/147", "detail": "Traceback (most recent call last):", "id": 43} -{"title": "报错:ImportError: cannot import name 'GENERATION_CONFIG_NAME' from 'transformers.utils'", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/149", "detail": "(mychatGLM) PS D:\\Users\\admin3\\zrh\\langchain-ChatGLM> python cli_demo.py", "id": 44} -{"title": "上传文件并加载知识库时,会不停地出现临时文件", "file": "2023-04-21.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/153", "detail": "环境:ubuntu 18.04", "id": 45} -{"title": "向知识库中添加文件后点击”上传文件并加载知识库“后Segmentation fault报错。", "file": "2023-04-23.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/161", "detail": "运行服务后的提示如下:", "id": 46} -{"title": "langchain-serve 集成", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/162", "detail": "Hey 我是来自 [langchain-serve](https://github.com/jina-ai/langchain-serve) 的dev!", "id": 47} -{"title": "大佬们,wsl的ubuntu怎么配置用cuda加速,装了运行后发现是cpu在跑", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/164", "detail": "大佬们,wsl的ubuntu怎么配置用cuda加速,装了运行后发现是cpu在跑", "id": 48} -{"title": "在github codespaces docker运行出错", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/165", "detail": "docker run -d --restart=always --name chatglm -p 7860:7860 -v /www/wwwroot/code/langchain-ChatGLM:/chatGLM chatglm", "id": 49} -{"title": "有计划接入Moss模型嘛", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/166", "detail": "后续会开展测试,目前主要在优化langchain部分效果,如果有兴趣也欢迎提PR", "id": 50} -{"title": "怎么实现 API 部署?", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/168", "detail": "利用 fastapi 实现 API 部署方式,具体怎么实现,有方法说明吗?", "id": 51} -{"title": " 'NoneType' object has no attribute 'message_types_by_name'报错", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/169", "detail": "_HISTOGRAMPROTO = DESCRIPTOR.message_types_by_name['HistogramProto']", "id": 52} -{"title": "能否指定自己训练的text2vector模型?", "file": "2023-04-25.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/172", "detail": "请问大佬:", "id": 53} -{"title": "关于项目支持的模型以及quantization_bit潜在的影响的问题", "file": "2023-04-26.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/176", "detail": "作者您好~", "id": 54} -{"title": "运行python3.9 api.py WARNING: You must pass the application as an import string to enable 'reload' or 'workers'.", "file": "2023-04-26.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/179", "detail": "api.py文件最下面改成这样试试:", "id": 55} -{"title": "ValidationError: 1 validation error for HuggingFaceEmbeddings model_kwargs extra fields not permitted (type=value_error.extra)", "file": "2023-04-26.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/180", "detail": "ValidationError: 1 validation error for HuggingFaceEmbeddings", "id": 56} -{"title": "如果没有检索到相关性比较高的,回答“我不知道”", "file": "2023-04-26.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/181", "detail": "如果通过设计system_template,让模型在搜索到的文档都不太相关的情况下回答“我不知道”", "id": 57} -{"title": "请问如果不能联网,6B之类的文件从本地上传需要放到哪里", "file": "2023-04-26.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/182", "detail": "感谢大佬的项目,很有启发~", "id": 58} -{"title": "知识库问答--输入新的知识库名称是中文的话,会报error", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/184", "detail": "知识库问答--输入新的知识库名称是中文的话,会报error,选择要加载的知识库那里也不显示之前添加的知识库", "id": 59} -{"title": "现在能通过问题匹配的相似度值,来直接返回文档中的文段,而不经过模型吗?因为有些答案在文档中,模型自己回答,不能回答文档中的答案", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/186", "detail": "现在能通过问题匹配的相似度值,来直接返回文档中的文段,而不经过模型吗?因为有些答案在文档中,模型自己回答,不能回答文档中的答案。也就是说,提供向量检索回答+模型回答相结合的策略。如果相似度值高于一定数值,直接返回文档中的文本,没有高于就返回模型的回答或者不知道", "id": 60} -{"title": "TypeError: The type of ChatGLM.callback_manager differs from the new default value; if you wish to change the type of this field, please use a type annotation", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/188", "detail": "Mac 运行 python3 ./webui.py 报 TypeError: The type of ChatGLM.callback_manager differs from the new default value; if you wish to change the type of this field, please use a type annotation", "id": 61} -{"title": "Not Enough Memory", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/190", "detail": "运行命令行程序python cli_demo.py, 已经成功加载pdf文件, 报“DefaultCPUAllocator: not enough memory: you tried to allocate 458288380900 bytes”错误,请问哪里可以配置default memory", "id": 62} -{"title": "参与开发问题", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/191", "detail": "1.是否需要进专门的开发群", "id": 63} -{"title": "对话框中代码片段格式需改进", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/192", "detail": "最好能改进下输出代码片段的格式,目前输出的格式还不友好。", "id": 64} -{"title": "请问未来有可能支持belle吗", "file": "2023-04-28.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/195", "detail": "如题,谢谢大佬", "id": 65} -{"title": "TypeError: cannot unpack non-iterable NoneType object", "file": "2023-04-28.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/200", "detail": "When i tried to change the knowledge vector store through `init_knowledge_vector_store`, the error `TypeError: cannot unpack non-iterable NoneType object` came out.", "id": 66} -{"title": "生成结果", "file": "2023-04-28.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/202", "detail": "你好,想问一下langchain+chatglm-6B,找到相似匹配的prompt,是直接返回prompt对应的答案信息,还是chatglm-6B在此基础上自己优化答案?", "id": 67} -{"title": "在win、ubuntu下都出现这个错误:attributeerror: 't5forconditionalgeneration' object has no attribute 'stream_chat'", "file": "2023-04-29.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/207", "detail": "在win、ubuntu。下载完模型后,没办法修改代码以执行本地模型,每次都要重新输入路径; LLM 模型、Embedding 模型支持也都在官网下的,在其他项目(wenda)下可以使用", "id": 68} -{"title": "[FEATURE] knowledge_based_chatglm.py: renamed or missing?", "file": "2023-04-30.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/210", "detail": "Not found. Was it renamed? Or, is it missing? How can I get it?", "id": 69} -{"title": "sudo apt-get install -y nvidia-container-toolkit-base执行报错", "file": "2023-05-01.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/211", "detail": "**问题描述 / Problem Description**", "id": 70} -{"title": "效果不佳几乎答不上来", "file": "2023-05-01.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/212", "detail": "提供了50条问答的docx文件", "id": 71} -{"title": "有没有可能新增一个基于chatglm api调用的方式构建langchain", "file": "2023-05-02.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/218", "detail": "我有两台8G GPU/40G内存的服务器,一个台做成了chatglm的api ;想基于另外一台服务器部署langchain;网上好像没有类似的代码。", "id": 72} -{"title": "电脑是intel的集成显卡; 运行时告知我找不到nvcuda.dll,模型无法运行", "file": "2023-05-02.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/219", "detail": "您好,我的电脑是intel的集成显卡,不过CPU是i5-11400 @ 2.60GHz ,内存64G;", "id": 73} -{"title": "根据langchain官方的文档和使用模式,是否可以改Faiss为Elasticsearch?会需要做哪些额外调整?求解", "file": "2023-05-03.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/221", "detail": "本人新手小白,由于业务模式的原因(有一些自己的场景和优化),希望利用Elasticsearch做这个体系内部的检索机制,不知道是否可以替换,同时,还会涉及到哪些地方的改动?或者说可能会有哪些其他影响,希望作者和大佬们不吝赐教!", "id": 74} -{"title": "请问未来有可能支持t5吗", "file": "2023-05-04.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/224", "detail": "请问可能支持基於t5的模型吗?", "id": 75} -{"title": "[BUG] 内存溢出 / torch.cuda.OutOfMemoryError:", "file": "2023-05-04.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/229", "detail": "**问题描述 / Problem Description**", "id": 76} -{"title": "报错 No module named 'chatglm_llm'", "file": "2023-05-04.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/230", "detail": "明明已经安装了包,却在python里吊不出来", "id": 77} -{"title": "能出一个api部署的描述文档吗", "file": "2023-05-04.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/233", "detail": "**功能描述 / Feature Description**", "id": 78} -{"title": "使用docs/API.md 出错", "file": "2023-05-04.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/234", "detail": "使用API.md文档2种方法,出错", "id": 79} -{"title": "加载pdf文档报错?", "file": "2023-05-05.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/238", "detail": "ew one with MEAN pooling.", "id": 80} -{"title": "上传的本地知识文件后再次上传不能显示,只显示成功了一个,别的上传成功后再次刷新就没了", "file": "2023-05-05.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/239", "detail": "您好,项目有很大启发,感谢~", "id": 81} -{"title": "创建了新的虚拟环境,安装了相关包,并且自动下载了相关的模型,但是仍旧出现:OSError: Unable to load weights from pytorch checkpoint file for", "file": "2023-05-05.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/240", "detail": "![78ac8e663fdc312d0e9d78da95925c4](https://user-images.githubusercontent.com/34124260/236378728-9ea4424f-0f7f-4013-9d33-820b723de321.png)", "id": 82} -{"title": "[BUG] 数据加载不进来", "file": "2023-05-05.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/243", "detail": "使用的.txt格式,utf-8编码,报以下错误", "id": 83} -{"title": "不能读取pdf", "file": "2023-05-05.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/244", "detail": "请问是webui还是cli_demo", "id": 84} -{"title": "本地txt文件有500M,加载的时候很慢,如何提高速度?", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/251", "detail": "![yayRzxSYHP](https://user-images.githubusercontent.com/109277248/236592902-f5ab338d-c1e9-43dc-ae16-9df2cd3c1378.jpg)", "id": 85} -{"title": "[BUG] gradio上传知识库后刷新之后 知识库就不见了 只有重启才能看到之前的上传的知识库", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/253", "detail": "gradio上传知识库后刷新之后 知识库就不见了 只有重启才能看到之前的上传的知识库", "id": 86} -{"title": "[FEATURE] 可以支持 OpenAI 的模型嘛?比如 GPT-3、GPT-3.5、GPT-4;embedding 增加 text-embedding-ada-002", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/254", "detail": "**功能描述 / Feature Description**", "id": 87} -{"title": "[FEATURE] 能否增加对于milvus向量数据库的支持 / Concise description of the feature", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/256", "detail": "**功能描述 / Feature Description**", "id": 88} -{"title": "CPU和GPU上跑,除了速度有区别,准确率效果回答上有区别吗?", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/259", "detail": "理论上没有区别", "id": 89} -{"title": "m1,请问在生成回答时怎么看是否使用了mps or cpu?", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/260", "detail": "m1,请问在生成回答时怎么看是否使用了mps or cpu?", "id": 90} -{"title": "知识库一刷新就没了", "file": "2023-05-07.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/263", "detail": "知识库上传后刷新就没了", "id": 91} -{"title": "本地部署报没有模型", "file": "2023-05-07.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/267", "detail": "建议在下载llm和embedding模型至本地后在configs/model_config中写入模型本地存储路径后再运行", "id": 92} -{"title": "[BUG] python3: can't open file 'webui.py': [Errno 2] No such file or directory", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/269", "detail": "**问题描述 / Problem Description**", "id": 93} -{"title": "模块缺失提示", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/271", "detail": "因为已有自己使用的docker环境,直接启动webui.py,提示", "id": 94} -{"title": "运行api.py后,执行curl -X POST \"http://127.0.0.1:7861\" 报错?", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/272", "detail": "执行curl -X POST \"http://127.0.0.1:7861\" \\ -H 'Content-Type: application/json' \\ -d '{\"prompt\": \"你好\", \"history\": []}',报错怎么解决", "id": 95} -{"title": "[BUG] colab安装requirements提示protobuf版本问题?", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/273", "detail": "pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.", "id": 96} -{"title": "请问项目里面向量相似度使用了什么方法计算呀?", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/275", "detail": "基本按照langchain里的FAISS.similarity_search_with_score_by_vector实现", "id": 97} -{"title": "[BUG] 安装detectron2后,pdf无法加载", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/276", "detail": "**问题描述 / Problem Description**", "id": 98} -{"title": "[BUG] 使用ChatYuan-V2模型无法流式输出,会报错", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/277", "detail": "一方面好像是ChatYuan本身不支持stream_chat,有人在clueai那边提了issue他们说还没开发,所以估计这个attribute调不起来;但是另一方面看报错好像是T5模型本身就不是decoder-only模型,所以不能流式输出吧(个人理解)", "id": 99} -{"title": "[BUG] 无法加载text2vec模型", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/278", "detail": "**问题描述 / Problem Description**", "id": 100} -{"title": "请问能否增加网络搜索功能", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/281", "detail": "请问能否增加网络搜索功能", "id": 101} -{"title": "[FEATURE] 结构化数据sql、excel、csv啥时会支持呐。", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/283", "detail": "**功能描述 / Feature Description**", "id": 102} -{"title": "TypeError: ChatGLM._call() got an unexpected keyword argument 'stop'", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/284", "detail": "No sentence-transformers model found with name D:\\DevProject\\langchain-ChatGLM\\GanymedeNil\\text2vec-large-chinese. Creating a new one with MEAN pooling.", "id": 103} -{"title": "关于api.py的一些bug和设计逻辑问题?", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/285", "detail": "首先冒昧的问一下,这个api.py,开发者大佬们是在自己电脑上测试后确实没问题吗?", "id": 104} -{"title": "有没有租用的算力平台上,运行api.py后,浏览器http://localhost:7861/报错", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/287", "detail": "是不是租用的gpu平台上都会出现这个问题???", "id": 105} -{"title": "请问一下项目中有用到文档段落切割方法吗?", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/288", "detail": "text_load中的文档切割方法用上了吗?在代码中看好像没有用到?", "id": 106} -{"title": "报错 raise ValueError(f\"Knowledge base {knowledge_base_id} not found\") ValueError: Knowledge base ./vector_store not found", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/289", "detail": "File \"/root/autodl-tmp/chatglm/langchain-ChatGLM-master/api.py\", line 183, in chat", "id": 107} -{"title": "能接入vicuna模型吗", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/290", "detail": "目前本地已经有了vicuna模型能直接接入吗?", "id": 108} -{"title": "[BUG] 提问公式相关问题大概率爆显存", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/291", "detail": "**问题描述 / Problem Description**", "id": 109} -{"title": "安装pycocotools失败,找了好多方法都不能解决。", "file": "2023-05-10.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/292", "detail": "**问题描述 / Problem Description**", "id": 110} -{"title": "使用requirements安装,PyTorch安装的是CPU版本", "file": "2023-05-10.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/294", "detail": "如题目,使用requirements安装,PyTorch安装的是CPU版本,运行程序的时候,也是使用CPU在工作。", "id": 111} -{"title": "能不能给一个毛坯服务器的部署教程", "file": "2023-05-10.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/298", "detail": "“开发部署”你当成服务器的部署教程用就行了。", "id": 112} -{"title": " Error(s) in loading state_dict for ChatGLMForConditionalGeneration:", "file": "2023-05-10.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/299", "detail": "运行中出现的问题,7860的端口页面显示不出来,求助。", "id": 113} -{"title": "ChatYuan-large-v2模型加载失败", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/300", "detail": "**实际结果 / Actual Result**", "id": 114} -{"title": "新增摘要功能", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/303", "detail": "你好,后续会考虑新增对长文本信息进行推理和语音理解功能吗?比如生成摘要", "id": 115} -{"title": "[BUG] pip install -r requirements.txt 出错", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/304", "detail": "pip install langchain -i https://pypi.org/simple", "id": 116} -{"title": "[BUG] 上传知识库文件报错", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/305", "detail": "![19621e29eaa547d01213bee53d81e6a](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/84606552/7f6ceb46-e494-4b0e-939c-23b585a6d9d8)", "id": 117} -{"title": "[BUG] AssertionError: Component with id 41 not a valid input component.", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/306", "detail": "**问题描述 / Problem Description**", "id": 118} -{"title": "[BUG] CUDA out of memory with container deployment", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/310", "detail": "**问题描述 / Problem Description**", "id": 119} -{"title": "[FEATURE] 增加微调训练功能", "file": "2023-05-11.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/311", "detail": "**功能描述 / Feature Description**", "id": 120} -{"title": "如何使用多卡部署,多个gpu", "file": "2023-05-11.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/315", "detail": "机器上有多个gpu,如何全使用了", "id": 121} -{"title": "请问这个知识库问答,和chatglm的关系是什么", "file": "2023-05-11.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/319", "detail": "这个知识库问答,哪部分关联到了chatglm,是不是没有这个chatglm,知识库问答也可单单拎出来", "id": 122} -{"title": "[BUG] 运行的时候报错ImportError: libcudnn.so.8: cannot open shared object file: No such file or directory", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/324", "detail": "**问题描述 / Problem Description**raceback (most recent call last):", "id": 123} -{"title": "webui启动成功,但有报错", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/325", "detail": "**问题描述 / Problem Description**", "id": 124} -{"title": "切换MOSS的时候报错", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/327", "detail": "danshi但是发布的源码中,", "id": 125} -{"title": "vicuna模型是否能接入?", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/328", "detail": "您好!关于MOSS模型和vicuna模型,都是AutoModelForCausalLM来加载模型的,但是稍作更改(模型路径这些)会报这个错误。这个错误的造成是什么", "id": 126} -{"title": "你好,请问一下在阿里云CPU服务器上跑可以吗?可以的话比较理想的cpu配置是什么?", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/330", "detail": "你好,请问一下在阿里云CPU服务器上跑可以吗?可以的话比较理想的cpu配置是什么?", "id": 127} -{"title": "你好,请问8核32g的CPU可以跑多轮对话吗?", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/331", "detail": "什么样的cpu配置比较好呢?我目前想部署CPU下的多轮对话?", "id": 128} -{"title": "[BUG] 聊天内容输入超过10000个字符系统出现错误", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/332", "detail": "聊天内容输入超过10000个字符系统出现错误,如下图所示:", "id": 129} -{"title": "能增加API的多用户访问接口部署吗?", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/333", "detail": "默认部署程序仅支持单用户访问,多用户则需要排队访问。测试过相关的几个Github多用户工程,但是其中一些仍然不满足要求。本节将系统介绍如何实现多用户同时访问ChatGLM的部署接口,包括http、websocket(流式输出,stream)和web页面等方式,主要目录如下所示。", "id": 130} -{"title": "多卡部署", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/334", "detail": "用单机多卡或多机多卡,fastapi部署模型,怎样提高并发", "id": 131} -{"title": "WEBUI能否指定知识库目录?", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/335", "detail": "**功能描述 / Feature Description**", "id": 132} -{"title": "[BUG] Cannot read properties of undefined (reading 'error')", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/336", "detail": "**问题描述 / Problem Description**", "id": 133} -{"title": "[BUG] 1 validation error for HuggingFaceEmbeddings model_kwargs extra fields not permitted.", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/337", "detail": "模型加载到 100% 后出现问题:", "id": 134} -{"title": "上传知识库需要重启能不能修复一下", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/338", "detail": "挺严重的这个问题", "id": 135} -{"title": "[BUG] 4块v100卡爆显存,在LLM会话模式也一样", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/339", "detail": "**问题描述 / Problem Description**", "id": 136} -{"title": "针对上传的文件配置不同的TextSpliter", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/341", "detail": "1. 目前的ChineseTextSpliter切分对英文尤其是代码文件不友好,而且限制固定长度;导致对话结果不如人意", "id": 137} -{"title": "[FEATURE] 未来可增加Bloom系列模型吗?根据甲骨易的测试,这系列中文评测效果不错", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/346", "detail": "**功能描述 / Feature Description**", "id": 138} -{"title": "[BUG] v0.1.12打包镜像后启动webui.py失败 / Concise description of the issue", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/347", "detail": "**问题描述 / Problem Description**", "id": 139} -{"title": "切换MOSS模型时报错", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/349", "detail": "昨天问了下,说是transformers版本不对,需要4.30.0,发现没有这个版本,今天更新到4.29.1,依旧报错,错误如下", "id": 140} -{"title": "[BUG] pdf文档加载失败", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/350", "detail": "**问题描述 / Problem Description**", "id": 141} -{"title": "建议可以在后期增强一波注释,这样也有助于更多人跟进提PR", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/351", "detail": "知道作者和团队在疯狂更新审查代码,只是建议后续稳定后可以把核心代码进行一些注释的补充,从而能帮助更多人了解各个模块作者的思路从而提出更好的优化。", "id": 142} -{"title": "[FEATURE] MOSS 量化版支援", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/353", "detail": "**功能描述 / Feature Description**", "id": 143} -{"title": "[BUG] moss模型无法加载", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/356", "detail": "**问题描述 / Problem Description**", "id": 144} -{"title": "[BUG] load_doc_qa.py 中的 load_file 函数有bug", "file": "2023-05-14.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/358", "detail": "原函数为:", "id": 145} -{"title": "[FEATURE] API模式,知识库加载优化", "file": "2023-05-14.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/359", "detail": "如题,当前版本,每次调用本地知识库接口,都将加载一次知识库,是否有更好的方式?", "id": 146} -{"title": "运行Python api.py脚本后端部署后,怎么使用curl命令调用?", "file": "2023-05-15.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/361", "detail": "也就是说,我现在想做个对话机器人,想和公司的前后端联调?怎么与前后端相互调用呢?可私信,有偿解答!!!", "id": 147} -{"title": "上传知识库需要重启能不能修复一下", "file": "2023-05-15.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/363", "detail": "上传知识库需要重启能不能修复一下", "id": 148} -{"title": "[BUG] pip install -r requirements.txt -i https://pypi.tuna.tsinghua.edu.cn/simple", "file": "2023-05-15.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/364", "detail": "我的python是3.8.5的", "id": 149} -{"title": "pip install gradio 报错", "file": "2023-05-15.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/367", "detail": "大佬帮我一下", "id": 150} -{"title": "[BUG] pip install gradio 一直卡不动", "file": "2023-05-15.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/369", "detail": "![aba82742dd9d4d242181662eb5027a7](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/84606552/cd9600d9-f6e7-46b7-b1be-30ed8b99f76b)", "id": 151} -{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/370", "detail": "初次加载本地知识库成功,但提问后,就无法重写加载本地知识库", "id": 152} -{"title": "[FEATURE] 简洁阐述功能 / Concise description of the feature", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/371", "detail": "**功能描述 / Feature Description**", "id": 153} -{"title": "在windows上,模型文件默认会安装到哪", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/372", "detail": "-------------------------------------------------------------------------------", "id": 154} -{"title": "[FEATURE] 兼顾对话管理", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/374", "detail": "如何在知识库检索的情况下,兼顾对话管理?", "id": 155} -{"title": "llm device: cpu embedding device: cpu", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/376", "detail": "**问题描述 / Problem Description**", "id": 156} -{"title": "[FEATURE] 简洁阐述功能 /文本文件的知识点之间使用什么分隔符可以分割?", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/377", "detail": "**功能描述 / Feature Description**", "id": 157} -{"title": "[BUG] 上传文件失败:PermissionError: [WinError 32] 另一个程序正在使用此文件,进程无法访问。", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/379", "detail": "**问题描述 / Problem Description**", "id": 158} -{"title": "[BUG] 执行python api.py 报错", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/383", "detail": "错误信息", "id": 159} -{"title": "model_kwargs extra fields not permitted (type=value_error.extra)", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/384", "detail": "大家好,请问这个有遇到的么,?", "id": 160} -{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-05-17.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/385", "detail": "执行的时候出现了ls1 = [ls[0]]", "id": 161} -{"title": "[FEATURE] 性能优化", "file": "2023-05-17.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/388", "detail": "**功能描述 / Feature Description**", "id": 162} -{"title": "[BUG] Moss模型问答,RuntimeError: probability tensor contains either inf, nan or element < 0", "file": "2023-05-17.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/390", "detail": "**问题描述 / Problem Description**", "id": 163} -{"title": "有没有人知道v100GPU的32G显存,会报错吗?支持V100GPU吗?", "file": "2023-05-17.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/392", "detail": "**问题描述 / Problem Description**", "id": 164} -{"title": "针对于编码问题比如'gbk' codec can't encode character '\\xab' in position 14: illegal multibyte sequence粗浅的解决方法", "file": "2023-05-17.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/397", "detail": "**功能描述 / Feature Description**", "id": 165} -{"title": "Could not import sentence_transformers python package. Please install it with `pip install sentence_transformers`.", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/400", "detail": "**问题描述 / Problem Description**", "id": 166} -{"title": "支持模型问答与检索问答", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/401", "detail": "不同的query,根据意图不一致,回答也应该不一样。", "id": 167} -{"title": "文本分割的时候,能不能按照txt文件的每行进行分割,也就是按照换行符号\\n进行分割???", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/403", "detail": "下面的代码应该怎么修改?", "id": 168} -{"title": "local_doc_qa/local_doc_chat 接口响应是串行", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/405", "detail": "**问题描述 / Problem Description**", "id": 169} -{"title": "为什么找到出处了,但是还是无法回答该问题?", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/406", "detail": "![图片](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/3349611/1fc81d61-2409-4330-9065-fdda1a27c86a)", "id": 170} -{"title": "请问下:知识库测试中的:添加单条内容,如果换成文本导入是是怎样的格式?我发现添加单条内容测试效果很好.", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/412", "detail": "我发现在知识库测试中`添加单条内容`,并且勾选`禁止内容分句入库`,即使 `不开启上下文关联`的测试效果都非常好.", "id": 171} -{"title": "[BUG] 无法配置知识库", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/413", "detail": "**问题描述 / Problem Description**", "id": 172} -{"title": "[BUG] 部署在阿里PAI平台的EAS上访问页面是白屏", "file": "2023-05-19.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/414", "detail": "**问题描述 / Problem Description**", "id": 173} -{"title": "API部署后调用/local_doc_qa/local_doc_chat 返回Knowledge base samples not found", "file": "2023-05-19.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/416", "detail": "入参", "id": 174} -{"title": "[FEATURE] 上传word另存为的txt文件报 'ascii' codec can't decode byte 0xb9 in position 6: ordinal not in range(128)", "file": "2023-05-20.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/421", "detail": "上传word另存为的txt文件报", "id": 175} -{"title": "创建保存的知识库刷新后没有出来,这个知识库是永久保存的吗?可以连外部的 向量知识库吗?", "file": "2023-05-21.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/422", "detail": "创建保存的知识库刷新后没有出来,这个知识库是永久保存的吗?可以连外部的 向量知识库吗?", "id": 176} -{"title": "[BUG] 用colab运行,无法加载模型,报错:'NoneType' object has no attribute 'message_types_by_name'", "file": "2023-05-21.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/423", "detail": "**问题描述 / Problem Description**", "id": 177} -{"title": "请问是否需要用到向量数据库?以及什么时候需要用到向量数据库?", "file": "2023-05-21.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/424", "detail": "目前用的是 text2vec , 请问是否需要用到向量数据库?以及什么时候需要用到向量数据库?", "id": 178} -{"title": "huggingface模型引用问题", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/427", "detail": "它最近似乎变成了一个Error?", "id": 179} -{"title": "你好,加载本地txt文件出现这个killed错误,TXT文件有100M左右大小。原因是?谢谢。", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/429", "detail": "\"929aca3b22b8cd74e997a87b61d241b\"", "id": 180} -{"title": "想请问一下,关于对本地知识的管理是如何管理?例如:通过http API接口添加数据 或者 删除某条数据", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/430", "detail": "例如:通过http API接口添加、删除、修改 某条数据。", "id": 181} -{"title": "[FEATURE] 双栏pdf识别问题", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/432", "detail": "试了一下模型,感觉对单栏pdf识别的准确性较高,但是由于使用的基本是ocr的技术,对一些双栏pdf论文识别出来有很多问题,请问有什么办法改善吗?", "id": 182} -{"title": "部署启动小问题,小弟初学求大佬解答", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/433", "detail": "1.python loader/image_loader.py时,提示ModuleNotFoundError: No module named 'configs',但是跑python webui.py还是还能跑", "id": 183} -{"title": "能否支持检测到目录下文档有增加而去增量加载文档,不影响前台对话,其实就是支持读写分离。如果能支持查询哪些文档向量化了,删除过时文档等就更好了,谢谢。", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/434", "detail": "**功能描述 / Feature Description**", "id": 184} -{"title": "[BUG] 简洁阐述问题 / windows 下cuda错误,请用https://github.com/Keith-Hon/bitsandbytes-windows.git", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/435", "detail": "pip install git+https://github.com/Keith-Hon/bitsandbytes-windows.git", "id": 185} -{"title": "[BUG] from commit 33bbb47, Required library version not found: libbitsandbytes_cuda121_nocublaslt.so. Maybe you need to compile it from source?", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/438", "detail": "**问题描述 / Problem Description**", "id": 186} -{"title": "[BUG] 简洁阐述问题 / Concise description of the issue上传60m的txt文件报错,显示超时,请问这个能上传的文件大小有限制吗", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/439", "detail": "ERROR 2023-05-23 11:13:09,627-1d: Timeout reached while detecting encoding for ./docs/GLM模型格式数据.txt", "id": 187} -{"title": "[BUG] TypeError: issubclass() arg 1 must be a class", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/440", "detail": "**问题描述**", "id": 188} -{"title": "执行python3 webui.py后,一直提示”模型未成功加载,请到页面左上角\"模型配置\"选项卡中重新选择后点击\"加载模型\"按钮“", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/441", "detail": "**问题描述 / Problem Description**", "id": 189} -{"title": "是否能提供网页文档得导入支持", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/444", "detail": "现在很多都是在线文档作为协作得工具,所以通过URL导入在线文档需求更大", "id": 190} -{"title": "[BUG] history 索引问题", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/445", "detail": "在比较对话框的history和模型chat function 中的history时, 发现并不匹配,在传入 llm._call 时,history用的索引是不是有点问题,导致上一轮对话的内容并不输入给模型。", "id": 191} -{"title": "[BUG] moss_llm没有实现", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/447", "detail": "有些方法没支持,如history_len", "id": 192} -{"title": "请问langchain-ChatGLM如何删除一条本地知识库的数据?", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/448", "detail": "例如:用户刚刚提交了一条错误的数据到本地知识库中了,现在如何在本地知识库从找到,并且对此删除。", "id": 193} -{"title": "[BUG] 简洁阐述问题 / UnboundLocalError: local variable 'resp' referenced before assignment", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/450", "detail": "在最新一版的代码中, 运行api.py 出现了以上错误(UnboundLocalError: local variable 'resp' referenced before assignment), 通过debug的方式观察到local_doc_qa.llm.generatorAnswer(prompt=question, history=history,streaming=True)可能不返回任何值。", "id": 194} -{"title": "请问有没有 PROMPT_TEMPLATE 能让模型不回答敏感问题", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/452", "detail": "## PROMPT_TEMPLATE问题", "id": 195} -{"title": "[BUG] 测试环境 Python 版本有误", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/456", "detail": "**问题描述 / Problem Description**", "id": 196} -{"title": "[BUG] webui 部署后样式不正确", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/458", "detail": "**问题描述 / Problem Description**", "id": 197} -{"title": "配置默认LLM模型的问题", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/459", "detail": "**问题描述 / Problem Description**", "id": 198} -{"title": "[FEATURE]是时候更新一下autoDL的镜像了", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/460", "detail": "如题,跑了下autoDL的镜像,发现是4.27号的,git pull新版本的代码功能+老的依赖环境,各种奇奇怪怪的问题。", "id": 199} -{"title": "[BUG] tag:0.1.13 以cpu模式下,想使用本地模型无法跑起来,各种路径参数问题", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/462", "detail": "-------------------------------------------------------------------------------", "id": 200} -{"title": "[BUG] 有没有同学遇到过这个错!!!加载本地txt文件出现这个killed错误,TXT文件有100M左右大小。", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/463", "detail": "运行cli_demo.py。是本地的txt文件太大了吗?100M左右。", "id": 201} -{"title": "API版本能否提供WEBSOCKET的流式接口", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/464", "detail": "webui 版本中,采用了WS的流式输出,整体感知反应很快", "id": 202} -{"title": "[BUG] 安装bug记录", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/465", "detail": "按照[install文档](https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/docs/INSTALL.md)安装的,", "id": 203} -{"title": "VUE的pnmp i执行失败的修复-用npm i命令即可", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/466", "detail": "感谢作者!非常棒的应用,用的很开心。", "id": 204} -{"title": "请教个问题,有没有人知道cuda11.4是否支持???", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/467", "detail": "请教个问题,有没有人知道cuda11.4是否支持???", "id": 205} -{"title": "请问有实现多轮问答中基于问题的搜索上下文关联么", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/468", "detail": "在基于知识库的多轮问答中,第一个问题讲述了一个主题,后续的问题描述没有包含这个主题的关键词,但又存在上下文的关联。如果用后续问题去搜索知识库有可能会搜索出无关的信息,从而导致大模型无法正确回答问题。请问这个项目要考虑这种情况吗?", "id": 206} -{"title": "[BUG] 内存不足的问题", "file": "2023-05-26.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/470", "detail": "我用了本地的chatglm-6b-int4模型,然后显示了内存不足(win10+32G内存+1080ti11G),一般需要多少内存才足够?这个bug应该如何解决?", "id": 207} -{"title": "[BUG] 纯内网环境安装pycocotools失败", "file": "2023-05-26.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/472", "detail": "**问题描述 / Problem Description**", "id": 208} -{"title": "[BUG] webui.py 重新加载模型会导致 KeyError", "file": "2023-05-26.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/473", "detail": "**问题描述 / Problem Description**", "id": 209} -{"title": "chatyuan无法使用", "file": "2023-05-26.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/475", "detail": "**问题描述 / Problem Description**", "id": 210} -{"title": "[BUG] 文本分割模型AliTextSplitter存在bug,会把“.”作为分割符", "file": "2023-05-26.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/476", "detail": "阿里达摩院的语义分割模型存在bug,默认会把\".”作为分割符进行分割而不管上下文语义。是否还有其他分割符则未知。建议的修改方案:把“.”统一替换为其他字符,分割后再替换回来。或者添加其他分割模型。", "id": 211} -{"title": "[BUG] RuntimeError: Error in faiss::FileIOReader::FileIOReader(const char*) a", "file": "2023-05-27.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/479", "detail": "**问题描述 / Problem Description**", "id": 212} -{"title": "[FEATURE] 安装,为什么conda create要额外指定路径 用-p ,而不是默认的/envs下面", "file": "2023-05-28.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/481", "detail": "##**功能描述 / Feature Description**", "id": 213} -{"title": "[小白求助] 通过Anaconda执行webui.py后,无法打开web链接", "file": "2023-05-28.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/485", "detail": "在执行webui.py命令后,http://0.0.0.0:7860复制到浏览器后无法打开,显示“无法访问此网站”。", "id": 214} -{"title": "[BUG] 使用 p-tuningv2后的模型,重新加载报错", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/486", "detail": "把p-tunningv2训练完后的相关文件放到了p-tunningv2文件夹下,勾选使用p-tuningv2点重新加载模型,控制台输错错误信息:", "id": 215} -{"title": "[小白求助] 服务器上执行webui.py后,在本地无法打开web链接", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/487", "detail": "此项目执行在xxx.xx.xxx.xxx服务器上,我在webui.py上的代码为 (demo", "id": 216} -{"title": "[FEATURE] 能不能支持VisualGLM-6B", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/488", "detail": "**功能描述 / Feature Description**", "id": 217} -{"title": "你好,问一下各位,后端api部署的时候,支持多用户同时问答吗???", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/489", "detail": "支持多用户的话,最多支持多少用户问答?根据硬件而定吧?", "id": 218} -{"title": "V100GPU显存占满,而利用率却为0,这是为什么?", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/491", "detail": "\"de45fe2b6cb76fa091b6e8f76a3de60\"", "id": 219} -{"title": "[求助] 如果在公司内部搭建产品知识库,使用INT-4模型,200人规模需要配置多少显存的服务器?", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/492", "detail": "如题,计划给公司搭一个在线知识库。", "id": 220} -{"title": "你好,请教个问题,目前问答回复需要20秒左右,如何提高速度?V10032G服务器。", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/493", "detail": "**问题描述 / Problem Description**", "id": 221} -{"title": "[FEATURE] 如何实现只匹配下文,而不要上文的结果", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/494", "detail": "在构建自己的知识库时,主要采用问答对的形式,那么也就是我需要的回答是在我的问题下面的内容,但是目前设置了chunk_size的值以后匹配的是上下文的内容,但我实际并不需要上文的。为了实现更完整的展示下面的答案,我只能调大chunk_size的值,但实际上上文的一半内容都是我不需要的。也就是扔了一半没用的东西给prompt,在faiss.py中我也没找到这块的一些描述,请问该如何进行修改呢?", "id": 222} -{"title": "你好,问一下,我调用api.py部署,为什么用ip加端口可以使用postman调用,而改为域名使用postman无法调用?", "file": "2023-05-30.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/497", "detail": "![5ufBSWxLyF](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/109277248/70e2fbac-5699-48d0-b0d1-3dc84fd042c2)", "id": 223} -{"title": "调用api.py中的stream_chat,返回source_documents中出现中文乱码。", "file": "2023-05-30.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/498", "detail": "-------------------------------------------------------------------------------", "id": 224} -{"title": "[BUG] 捉个虫,api.py中的stream_chat解析json问题", "file": "2023-05-30.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/501", "detail": "**问题描述 / Problem Description**", "id": 225} -{"title": "windows本地部署遇到了omp错误", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/502", "detail": "**问题描述 / Problem Description**", "id": 226} -{"title": "[BUG] bug14 ,\"POST /local_doc_qa/upload_file HTTP/1.1\" 422 Unprocessable Entity", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/503", "detail": "上传的文件报错,返回错误,api.py", "id": 227} -{"title": "你好,请教个问题,api.py部署的时候,如何改为多线程调用?谢谢", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/505", "detail": "目前的api.py脚本不支持多线程", "id": 228} -{"title": "你好,请教一下。api.py部署的时候,能不能提供给后端流失返回结果。", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/507", "detail": "curl -X 'POST' \\", "id": 229} -{"title": "流式输出,流式接口,使用server-sent events技术。", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/508", "detail": "想这样一样,https://blog.csdn.net/weixin_43228814/article/details/130063010", "id": 230} -{"title": "计划增加流式输出功能吗?ChatGLM模型通过api方式调用响应时间慢怎么破,Fastapi流式接口来解惑,能快速提升响应速度", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/509", "detail": "**问题描述 / Problem Description**", "id": 231} -{"title": "[BUG] 知识库上传时发生ERROR (could not open xxx for reading: No such file or directory)", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/510", "detail": "**问题描述 / Problem Description**", "id": 232} -{"title": "api.py脚本打算增加SSE流式输出吗?", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/511", "detail": "curl调用的时候可以检测第一个字,从而提升回复的体验", "id": 233} -{"title": "[BUG] 使用tornado实现webSocket,可以多个客户端同时连接,并且实现流式回复,但是多个客户端同时使用,答案就很乱,是模型不支持多线程吗", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/512", "detail": "import asyncio", "id": 234} -{"title": "支持 chinese_alpaca_plus_lora 吗 基于llama的", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/514", "detail": "支持 chinese_alpaca_plus_lora 吗 基于llama的,https://github.com/ymcui/Chinese-LLaMA-Alpaca这个项目的", "id": 235} -{"title": "[BUG] 现在能读图片的pdf了,但是文字的pdf反而读不了了,什么情况???", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/515", "detail": "**问题描述 / Problem Description**", "id": 236} -{"title": "在推理的过程中卡住不动,进程无法正常结束", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/516", "detail": "**问题描述 / Problem Description**", "id": 237} -{"title": "curl调用的时候,从第二轮开始,curl如何传参可以实现多轮对话?", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/517", "detail": "第一轮调用:", "id": 238} -{"title": "建议添加api.py部署后的日志管理功能?", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/518", "detail": "-------------------------------------------------------------------------------", "id": 239} -{"title": "有大佬知道,怎么多线程部署api.py脚本吗?", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/519", "detail": "api.py部署后,使用下面的请求,时间较慢,好像是单线程,如何改为多线程部署api.py:", "id": 240} -{"title": "[BUG] 上传文件到知识库 任何格式与内容都永远失败", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/520", "detail": "上传知识库的时候,传txt无法解析,就算是穿content/sample里的样例txt也无法解析,上传md、pdf等都无法加载,会持续性等待,等到了超过30分钟也不行。", "id": 241} -{"title": "关于prompt_template的问题", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/521", "detail": "请问这段prompt_template是什么意思,要怎么使用?可以给一个具体模板参考下吗?", "id": 242} -{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/522", "detail": "**问题描述 / Problem Description**", "id": 243} -{"title": "中文分词句号处理(关于表达金额之间的\".\")", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/523", "detail": "建议处理12.6亿元的这样的分词,最好别分成12 和6亿这样的,需要放到一起", "id": 244} -{"title": "ImportError: cannot import name 'inference' from 'paddle'", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/526", "detail": "在网上找了一圈,有说升级paddle的,我做了还是没有用,有说安装paddlepaddle的,我找了豆瓣的镜像源,但安装报错cannot detect archive format", "id": 245} -{"title": "[BUG] webscoket 接口串行问题(/local_doc_qa/stream-chat/{knowledge_base_id})", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/527", "detail": "**问题描述 / Problem Description**", "id": 246} -{"title": "[FEATURE] 刷新页面更新知识库列表", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/528", "detail": "**功能描述以及改进方案**", "id": 247} -{"title": "[BUG] 使用ptuning微调模型后,问答效果并不好", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/530", "detail": "### 未调用ptuning", "id": 248} -{"title": "[BUG] 多轮对话效果不佳", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/532", "detail": "在进行多轮对话的时候,无论设置的history_len是多少,效果都不好。事实上我将其设置成了最大值10,但在对话中,仍然无法实现多轮对话:", "id": 249} -{"title": "RuntimeError: MPS backend out of memory (MPS allocated: 18.00 GB, other allocations: 4.87 MB, max allowed: 18.13 GB)", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/533", "detail": "**问题描述**", "id": 250} -{"title": " 请大家重视这个issue!真正使用肯定是多用户并发问答,希望增加此功能!!!", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/534", "detail": "这得看你有多少显卡", "id": 251} -{"title": "在启动项目的时候如何使用到多张gpu啊?", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/535", "detail": "**在启动项目的时候如何使用到多张gpu啊?**", "id": 252} -{"title": " 使用流式输出的时候,curl调用的格式是什么?", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/536", "detail": "app.websocket(\"/local_doc_qa/stream-chat/{knowledge_base_id}\")(stream_chat)中的knowledge_base_id应该填什么???", "id": 253} -{"title": "使用本地 vicuna-7b模型启动错误", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/538", "detail": "环境: ubuntu 22.04 cuda 12.1 没有安装nccl,使用rtx2080与m60显卡并行计算", "id": 254} -{"title": "为什么会不调用GPU直接调用CPU呢", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/539", "detail": "我的阿里云配置是16G显存,用默认代码跑webui.py时提示", "id": 255} -{"title": "上传多个文件时会互相覆盖", "file": "2023-06-03.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/541", "detail": "1、在同一个知识库中上传多个文件时会互相覆盖,无法结合多个文档的知识,有大佬知道怎么解决吗?", "id": 256} -{"title": "[BUG] ‘gcc’不是内部或外部命令/LLM对话只能持续一轮", "file": "2023-06-03.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/542", "detail": "No compiled kernel found.", "id": 257} -{"title": "以API模式启动项目却没有知识库的接口列表?", "file": "2023-06-04.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/544", "detail": "请问如何获取知识库的接口列表?如果没有需要自行编写的话,可不可以提供相关的获取方式,感谢", "id": 258} -{"title": "程序以API模式启动的时候,如何才能让接口以stream模式被调用呢?", "file": "2023-06-05.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/546", "detail": "作者您好,我在以API模式进行程序启动后,我发现接口响应时间很长,怎么样才能让接口以stream模式被调用呢?我想实现像webui模式的回答那样", "id": 259} -{"title": "关于原文中表格转为文本后数据相关度问题。", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/547", "detail": "原文中表格数据转换为文本,以 (X-Y:值;...) 的格式每一行组织成一句话,但这样做后发现相关度较低,效果很差,有何好的方案吗?", "id": 260} -{"title": "启动后LLM和知识库问答模式均只有最后一轮记录", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/548", "detail": "拉取最新代码,问答时,每次页面只显示最后一次问答记录,需要修改什么参数才可以保留历史记录?", "id": 261} -{"title": "提供system message配置,以便于让回答不要超出知识库范围", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/549", "detail": "**功能描述 / Feature Description**", "id": 262} -{"title": "[BUG] 使用p-tunningv2报错", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/551", "detail": "按照readme的指示把p-tunningv2训练完后的文件放到了p-tunningv2文件夹下,勾选使用p-tuningv2点重新加载模型,控制台提示错误信息:", "id": 263} -{"title": "[BUG] 智障,这么多问题,也好意思放出来,浪费时间", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/553", "detail": "。。。", "id": 264} -{"title": "[FEATURE] 我看代码文件中有一个ali_text_splitter.py,为什么不用他这个文本分割器了?", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/554", "detail": "我看代码文件中有一个ali_text_splitter.py,为什么不用他这个文本分割器了?", "id": 265} -{"title": "加载文档函数报错", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/557", "detail": "def load_file(filepath, sentence_size=SENTENCE_SIZE):", "id": 266} -{"title": "参考指引安装docker后,运行cli_demo.py,提示killed", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/558", "detail": "root@b3d1bd08095c:/chatGLM# python3 cli_demo.py", "id": 267} -{"title": "注意:如果安装错误,注意这两个包的版本 wandb==0.11.0 protobuf==3.18.3", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/559", "detail": "Error1: 如果启动异常报错 `protobuf` 需要更新到 `protobuf==3.18.3 `", "id": 268} -{"title": "知识库对长文的知识相关度匹配不太理想有何优化方向", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/563", "detail": "我们可能录入一个文章有 1W 字,里面涉及这个文章主题的很多角度问题,我们针对他提问,他相关度匹配的内容和实际我们需要的答案相差很大怎么办。", "id": 269} -{"title": "使用stream-chat函数进行流式输出的时候,能使用curl调用吗?", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/565", "detail": "为什么下面这样调用会报错???", "id": 270} -{"title": "有大佬实践过 并行 或者 多线程 的部署方案吗?", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/566", "detail": "+1", "id": 271} -{"title": "多线程部署遇到问题?", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/567", "detail": "\"3d87bf74f0cf1a4820cc9e46b245859\"", "id": 272} -{"title": "[BUG] 用fastchat加载vicuna-13b模型进行知识库的问答有token的限制错误", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/569", "detail": "当我开启fastchat的vicuna-13b的api服务,然后config那里配置好(api本地测试过可以返回结果),然后知识库加载好之后(知识库大概有1000多个文档,用chatGLM可以正常推理),进行问答时出现token超过限制,就问了一句hello;", "id": 273} -{"title": "现在的添加知识库,文件多了总是报错,也不知道自己加载了哪些文件,报错后也不知道是全部失败还是一部分成功;希望能有个加载指定文件夹作为知识库的功能", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/574", "detail": "**功能描述 / Feature Description**", "id": 274} -{"title": "[BUG] moss模型本地加载报错", "file": "2023-06-08.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/577", "detail": "moss模型本地加载报错:", "id": 275} -{"title": "加载本地moss模型报错Can't instantiate abstract class MOSSLLM with abstract methods _history_len", "file": "2023-06-08.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/578", "detail": "(vicuna) ps@ps[13:56:20]:/data/chat/langchain-ChatGLM2/langchain-ChatGLM-0.1.13$ python webui.py --model-dir local_models --model moss --no-remote-model", "id": 276} -{"title": "[FEATURE] 能增加在前端页面控制prompt_template吗?或是能支持前端页面选择使用哪个prompt?", "file": "2023-06-08.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/579", "detail": "目前只能在config里修改一个prompt,想在多个不同场景切换比较麻烦", "id": 277} -{"title": "[BUG] streamlit ui的bug,在增加知识库时会报错", "file": "2023-06-08.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/580", "detail": "**问题描述 / Problem Description**", "id": 278} -{"title": "[FEATURE] webui/webui_st可以支持history吗?目前仅能一次对话", "file": "2023-06-08.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/581", "detail": "试了下webui和webui_st都不支持历史对话啊,只能对话一次,不能默认开启所有history吗?", "id": 279} -{"title": "启动python cli_demo.py --model chatglm-6b-int4-qe报错", "file": "2023-06-09.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/585", "detail": "下载好模型,和相关依赖环境,之间运行`python cli_demo.py --model chatglm-6b-int4-qe`报错了:", "id": 280} -{"title": "重新构建知识库报错", "file": "2023-06-09.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/586", "detail": "**问题描述 / Problem Description**", "id": 281} -{"title": "[FEATURE] 能否屏蔽paddle,我不需要OCR,效果差依赖环境还很复杂", "file": "2023-06-09.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/587", "detail": "希望能不依赖paddle", "id": 282} -{"title": "question :文档向量化这个可以自己手动实现么?", "file": "2023-06-09.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/589", "detail": "现有公司级数据500G+,需要使用这个功能,请问如何手动实现这个向量化,然后并加载", "id": 283} -{"title": "view前端能进行流式的返回吗??", "file": "2023-06-09.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/590", "detail": "view前端能进行流式的返回吗??", "id": 284} -{"title": "[BUG] Load parallel cpu kernel failed, using default cpu kernel code", "file": "2023-06-11.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/594", "detail": "**问题描述 / Problem Description**", "id": 285} -{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-06-11.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/595", "detail": "**问题描述 / Problem Description**", "id": 286} -{"title": "我在上传本地知识库时提示KeyError: 'name'错误,本地知识库都是.txt文件,文件数量大约是2000+。", "file": "2023-06-12.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/597", "detail": "\"KError\"", "id": 287} -{"title": "model_config.py中有vicuna-13b-hf模型的配置信息,但是好像还是不可用?", "file": "2023-06-12.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/600", "detail": "@dongyihua543", "id": 288} -{"title": "ImportError: Using SOCKS proxy, but the 'socksio' package is not installed. Make sure to install httpx using `pip install httpx[socks]`.", "file": "2023-06-12.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/605", "detail": "应该代理问题,但是尝试了好多方法都解决不了,", "id": 289} -{"title": "[BUG] similarity_search_with_score_by_vector在找不到匹配的情况下出错", "file": "2023-06-12.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/607", "detail": "在设置匹配阈值 VECTOR_SEARCH_SCORE_THRESHOLD 的情况下,vectorstore会返回空,此时上述处理函数会出错", "id": 290} -{"title": "[FEATURE] 请问如何搭建英文知识库呢", "file": "2023-06-12.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/609", "detail": "**功能描述 / Feature Description**", "id": 291} -{"title": "谁有vicuna权重?llama转换之后的", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/611", "detail": "**问题描述 / Problem Description**", "id": 292} -{"title": "[FEATURE] API能实现上传文件夹的功能么?", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/612", "detail": "用户懒得全选所有的文件,就想上传个文件夹,请问下API能实现这个功能么?", "id": 293} -{"title": "请问在多卡部署后,上传单个文件作为知识库,用的是单卡在生成向量还是多卡?", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/614", "detail": "目前我检测我本地多卡部署的,好像生成知识库向量的时候用的还是单卡", "id": 294} -{"title": "[BUG] python webui.py提示非法指令", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/615", "detail": "(/data/conda-langchain [root@chatglm langchain-ChatGLM]# python webui.py", "id": 295} -{"title": "知识库文件跨行切分问题", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/616", "detail": "我的知识库文件txt文件,是一行一条知识,用\\n分行。", "id": 296} -{"title": "[FEATURE] bing搜索问答有流式的API么?", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/617", "detail": "web端是有这个bing搜索回答,但api接口没有发现,大佬能给个提示么?", "id": 297} -{"title": "希望出一个macos m2的安装教程", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/620", "detail": "mac m2安装,模型加载成功了,知识库文件也上传成功了,但是一问答就会报错,报错内容如下", "id": 298} -{"title": "为【出处】提供高亮显示", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/621", "detail": "具体出处里面,对相关的内容高亮显示,不包含前后文。", "id": 299} -{"title": "[BUG] CPU运行cli_demo.py,不回答,hang住", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/622", "detail": "没有GPU;32G内存的ubuntu机器。", "id": 300} -{"title": "关于删除知识库里面的文档后,LLM知识库对话的时候还是会返回该被删除文档的内容", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/623", "detail": "如题,在vue前端成功执行删除知识库里面文档A.txt后,未能也在faiss索引中也删除该文档,LLM还是会返回这个A.txt的内容,并且以A.txt为出处,未能达到删除的效果", "id": 301} -{"title": "[BUG] 调用知识库进行问答,显存会一直叠加", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/625", "detail": "14G的显存,调用的chatglm-6b-int8模型,进行知识库问答时,最多问答四次就会爆显存了,观察了一下显存使用情况,每一次使用就会增加一次显存,请问这样是正常的吗?是否有什么配置需要开启可以解决这个问题?例如进行一次知识库问答清空上次问题的显存?", "id": 302} -{"title": "[BUG] web页面 重新构建数据库 失败,导致 原来的上传的数据库都没了", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/626", "detail": "web页面 重新构建数据库 失败,导致 原来的上传的数据库都没了", "id": 303} -{"title": "在CPU上运行webui.py报错Tensor on device cpu is not on the expected device meta!", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/627", "detail": "在CPU上运行python webui.py能启动,但最后有:RuntimeError: Tensor on device cpu is not on the expected device meta!", "id": 304} -{"title": "OSError: [WinError 1114] 动态链接库(DLL)初始化例程失败。 Error loading \"E:\\xxx\\envs\\langchain\\lib\\site-packages\\torch\\lib\\caffe2_nvrtc.dll\" or one of its dependencies.哪位大佬知道如何解决吗?", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/629", "detail": "**问题描述 / Problem Description**", "id": 305} -{"title": "[BUG] WEBUI删除知识库文档,会导致知识库问答失败", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/632", "detail": "如题,从知识库已有文件中选择要删除的文件,点击删除后,在问答框输入内容回车报错", "id": 306} -{"title": "更新后的版本中,删除知识库中的文件,再提问出现error错误", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/634", "detail": "针对更新版本,识别到一个问题,过程如下:", "id": 307} -{"title": "我配置好了环境,想要实现本地知识库的问答?可是它返回给我的", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/637", "detail": "没有总结,只有相关度的回复,但是我看演示里面表现的,回复是可以实现总结的,我去查询代码", "id": 308} -{"title": "[BUG] NPM run dev can not successfully start the VUE frontend", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/638", "detail": "**问题描述 / Problem Description**", "id": 309} -{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/639", "detail": "**问题描述 / Problem Description**", "id": 310} -{"title": "提一个模型加载的bug,我在截图中修复了,你们有空可以看一下。", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/642", "detail": "![model_load_bug](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/59411575/4432adc4-ccdd-45d9-aafc-5f2d1963403b)", "id": 311} -{"title": "[求助]关于设置embedding model路径的问题", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/643", "detail": "如题,我之前成功跑起来过一次,但因环境丢失重新配置 再运行webui就总是报错", "id": 312} -{"title": "Lora微调后的模型可以直接使用吗", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/646", "detail": "看model_config.py里是有USE_LORA这个参数的,但是在cli_demo.py和webui.py这两个里面都没有用到,实际测试下来模型没有微调的效果,想问问现在这个功能实现了吗", "id": 313} -{"title": "write_check_file在tmp_files目录下生成的load_file.txt是否需要一直保留,占用空间很大,在建完索引后能否删除", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/647", "detail": "**功能描述 / Feature Description**", "id": 314} -{"title": "[BUG] /local_doc_qa/list_files?knowledge_base_id=test删除知识库bug", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/649", "detail": "1.新建test知识库并上传文件(在vue前端完成并检查后端发现确实生成了test文件夹以及下面的content和vec_store", "id": 315} -{"title": "[BUG] vue webui无法加载知识库", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/650", "detail": "拉取了最新的代码,分别运行了后端api和前端web,点击知识库,始终只能显示simple,无法加载知识库", "id": 316} -{"title": "不能本地加载moss模型吗?", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/652", "detail": "手动下载模型设置local_model_path路径依旧提示缺少文件,该如何正确配置?", "id": 317} -{"title": "macos m2 pro docker 安装失败", "file": "2023-06-17.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/654", "detail": "macos m2 pro docker 安装失败", "id": 318} -{"title": " [BUG] mac m1 pro 运行提示 zsh: segmentation fault", "file": "2023-06-17.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/655", "detail": "运行: python webui.py", "id": 319} -{"title": "安装 requirements 报错", "file": "2023-06-17.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/656", "detail": "(langchainchatglm) D:\\github\\langchain-ChatGLM>pip install -r requirements.txt -i https://pypi.tuna.tsinghua.edu.cn/simple/", "id": 320} -{"title": "[BUG] AssertionError", "file": "2023-06-17.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/658", "detail": "**问题描述 / Problem Description**", "id": 321} -{"title": "[FEATURE] 支持AMD win10 本地部署吗?", "file": "2023-06-18.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/660", "detail": "**功能描述 / Feature Description**", "id": 322} +{"title": "效果如何优化", "file": "2023-04-04.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/14", "detail": "如图所示,将该项目的README.md和该项目结合后,回答效果并不理想,请问可以从哪些方面进行优化", "id": 0} +{"title": "怎么让模型严格根据检索的数据进行回答,减少胡说八道的回答呢", "file": "2023-04-04.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/15", "detail": "举个例子:", "id": 1} +{"title": "When I try to run the `python knowledge_based_chatglm.py`, I got this error in macOS(M1 Max, OS 13.2)", "file": "2023-04-07.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/32", "detail": "```python", "id": 2} +{"title": "萌新求教大佬怎么改成AMD显卡或者CPU?", "file": "2023-04-10.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/48", "detail": "把.cuda()去掉就行", "id": 3} +{"title": "输出answer的时间很长,是否可以把文本向量化的部分提前做好存储起来?", "file": "2023-04-10.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/50", "detail": "GPU:4090 24G显存", "id": 4} +{"title": "报错Use `repo_type` argument if needed.", "file": "2023-04-11.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/57", "detail": "Traceback (most recent call last):", "id": 5} +{"title": "无法打开gradio的页面", "file": "2023-04-11.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/58", "detail": "$ python webui.py", "id": 6} +{"title": "支持word,那word里面的图片正常显示吗?", "file": "2023-04-12.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/60", "detail": "如题,刚刚从隔壁转过来的,想先了解下", "id": 7} +{"title": "detectron2 is not installed. Cannot use the hi_res partitioning strategy. Falling back to partitioning with the fast strategy.", "file": "2023-04-12.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/63", "detail": "能够正常的跑起来,在加载content文件夹中的文件时,每加载一个文件都会提示:", "id": 8} +{"title": "cpu上运行webui,step3 asking时报错", "file": "2023-04-12.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/66", "detail": "web运行,文件加载都正常,asking时报错", "id": 9} +{"title": "建议弄一个插件系统", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/67", "detail": "如题弄成stable-diffusion-webui那种能装插件,再开一个存储库给使用者或插件开发,存储或下载插件。", "id": 10} +{"title": "请教加载模型出错!?", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/75", "detail": "AttributeError: module 'transformers_modules.chatglm-6b.configuration_chatglm' has no attribute 'ChatGLMConfig 怎么解决呀", "id": 11} +{"title": "从本地知识检索内容的时候,是否可以设置相似度阈值,小于这个阈值的内容不返回,即使会小于设置的VECTOR_SEARCH_TOP_K参数呢?谢谢大佬", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/76", "detail": "比如 问一些 你好/你是谁 等一些跟本地知识库无关的问题", "id": 12} +{"title": "如何改成多卡推理?", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/77", "detail": "+1", "id": 13} +{"title": "能否弄个懒人包,可以一键体验?", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/78", "detail": "能否弄个懒人包,可以一键体验?", "id": 14} +{"title": "连续问问题会导致崩溃", "file": "2023-04-13.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/79", "detail": "看上去不是爆内存的问题,连续问问题后,会出现如下报错", "id": 15} +{"title": "AttributeError: 'NoneType' object has no attribute 'as_retriever'", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/86", "detail": "环境:windows 11, anaconda/python 3.8", "id": 16} +{"title": "FileNotFoundError: Could not find module 'nvcuda.dll' (or one of its dependencies). Try using the full path with constructor syntax.", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/87", "detail": "请检查一下cuda或cudnn是否存在安装问题", "id": 17} +{"title": "加载txt文件失败?", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/89", "detail": "![JppHrGOWFa](https://user-images.githubusercontent.com/109277248/232009383-bf7c46d1-a01e-4e0a-9de6-5b5ed3e36158.jpg)", "id": 18} +{"title": "NameError: name 'chatglm' is not defined", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/90", "detail": "This share link expires in 72 hours. For free permanent hosting and GPU upgrades (NEW!), check out Spaces: https://huggingface.co/spaces", "id": 19} +{"title": "打不开地址?", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/91", "detail": "报错数据如下:", "id": 20} +{"title": "加载md文件出错", "file": "2023-04-14.00", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/98", "detail": "运行 webui.py后能访问页面,上传一个md文件后,日志中有错误。等待后能加载完成,提示可以提问了,但提问没反应,日志中有错误。 具体日志如下。", "id": 21} +{"title": "建议增加获取在线知识的能力", "file": "2023-04-15.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/101", "detail": "建议增加获取在线知识的能力", "id": 22} +{"title": "txt 未能成功加载", "file": "2023-04-15.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/103", "detail": "hinese. Creating a new one with MEAN pooling.", "id": 23} +{"title": "pdf加载失败", "file": "2023-04-15.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/105", "detail": "e:\\a.txt加载成功了,e:\\a.pdf加载就失败,pdf文件里面前面几页是图片,后面都是文字,加载失败没有报更多错误,请问该怎么排查?", "id": 24} +{"title": "一直停在文本加载处", "file": "2023-04-15.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/108", "detail": "一直停在文本加载处", "id": 25} +{"title": " File \"/root/.cache/huggingface/modules/transformers_modules/chatglm-6b/modeling_chatglm.py\", line 440, in forward new_tensor_shape = mixed_raw_layer.size()[:-1] + ( TypeError: torch.Size() takes an iterable of 'int' (item 2 is 'float')", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/113", "detail": "按照最新的代码,发现", "id": 26} +{"title": "后续会提供前后端分离的功能吗?", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/114", "detail": "类似这种https://github.com/lm-sys/FastChat/tree/main/fastchat/serve", "id": 27} +{"title": "安装依赖报错", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/115", "detail": "(test) C:\\Users\\linh\\Desktop\\langchain-ChatGLM-master>pip install -r requirements.txt", "id": 28} +{"title": "问特定问题会出现爆显存", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/117", "detail": "正常提问没问题。", "id": 29} +{"title": "Expecting value: line 1 column 1 (char 0)", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/118", "detail": "运行后 第一步加载配置一直报错:", "id": 30} +{"title": "embedding https://huggingface.co/GanymedeNil/text2vec-large-chinese/tree/main是免费的,效果比对openai的如何?", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/119", "detail": "-------------------------------------------------------------------------------", "id": 31} +{"title": "这是什么错误,在Colab上运行的。", "file": "2023-04-17.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/120", "detail": "libcuda.so.1: cannot open shared object file: No such file or directory", "id": 32} +{"title": "只想用自己的lora微调后的模型进行对话,不想加载任何本地文档,该如何调整?", "file": "2023-04-18.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/121", "detail": "能出一个单独的教程吗", "id": 33} +{"title": "租的gpu,Running on local URL: http://0.0.0.0:7860 To create a public link, set `share=True` in `launch()`. 浏览器上访问不了???", "file": "2023-04-18.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/122", "detail": "(chatglm20230401) root@autodl-container-e82d11963c-10ece0d7:~/autodl-tmp/chatglm/langchain-ChatGLM-20230418# python3.9 webui.py", "id": 34} +{"title": "本地部署中的报错请教", "file": "2023-04-18.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/124", "detail": "您好,在本地运行langchain-ChatGLM过程中,环境及依赖的包都已经满足条件,但是运行webui.py,报错如下(运行cli_demo.py报错类似),请问是哪里出了错呢?盼望您的回复,谢谢!", "id": 35} +{"title": "报错。The dtype of attention mask (torch.int64) is not bool", "file": "2023-04-18.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/131", "detail": "The dtype of attention mask (torch.int64) is not bool", "id": 36} +{"title": "[求助] pip install -r requirements.txt 的时候出现以下报错。。。有大佬帮忙看看怎么搞么,下的release里面的包", "file": "2023-04-18.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/134", "detail": "$ pip install -r requirements.txt", "id": 37} +{"title": "如何提升根据问题搜索到对应知识的准确率", "file": "2023-04-19.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/136", "detail": "外链知识库最大的问题在于问题是短文本,知识是中长文本。如何根据问题精准的搜索到对应的知识是个最大的问题。这类本地化项目不像百度,由无数的网页,基本上每个问题都可以找到对应的页面。", "id": 38} +{"title": "是否可以增加向量召回的阈值设定,有些召回内容相关性太低,导致模型胡言乱语", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/140", "detail": "如题", "id": 39} +{"title": "输入长度问题", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/141", "detail": "感谢作者支持ptuning微调模型。", "id": 40} +{"title": "已有部署好的chatGLM-6b,如何通过接口接入?", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/144", "detail": "已有部署好的chatGLM-6b,如何通过接口接入,而不是重新加载一个模型;", "id": 41} +{"title": "执行web_demo.py后,显示Killed,就退出了,是不是配置不足呢?", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/146", "detail": "![图片](https://user-images.githubusercontent.com/26102866/233256425-c7aab999-11d7-4de9-867b-23ef18d519e4.png)", "id": 42} +{"title": "执行python cli_demo1.py", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/147", "detail": "Traceback (most recent call last):", "id": 43} +{"title": "报错:ImportError: cannot import name 'GENERATION_CONFIG_NAME' from 'transformers.utils'", "file": "2023-04-20.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/149", "detail": "(mychatGLM) PS D:\\Users\\admin3\\zrh\\langchain-ChatGLM> python cli_demo.py", "id": 44} +{"title": "上传文件并加载知识库时,会不停地出现临时文件", "file": "2023-04-21.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/153", "detail": "环境:ubuntu 18.04", "id": 45} +{"title": "向知识库中添加文件后点击”上传文件并加载知识库“后Segmentation fault报错。", "file": "2023-04-23.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/161", "detail": "运行服务后的提示如下:", "id": 46} +{"title": "langchain-serve 集成", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/162", "detail": "Hey 我是来自 [langchain-serve](https://github.com/jina-ai/langchain-serve) 的dev!", "id": 47} +{"title": "大佬们,wsl的ubuntu怎么配置用cuda加速,装了运行后发现是cpu在跑", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/164", "detail": "大佬们,wsl的ubuntu怎么配置用cuda加速,装了运行后发现是cpu在跑", "id": 48} +{"title": "在github codespaces docker运行出错", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/165", "detail": "docker run -d --restart=always --name chatglm -p 7860:7860 -v /www/wwwroot/code/langchain-ChatGLM:/chatGLM chatglm", "id": 49} +{"title": "有计划接入Moss模型嘛", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/166", "detail": "后续会开展测试,目前主要在优化langchain部分效果,如果有兴趣也欢迎提PR", "id": 50} +{"title": "怎么实现 API 部署?", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/168", "detail": "利用 fastapi 实现 API 部署方式,具体怎么实现,有方法说明吗?", "id": 51} +{"title": " 'NoneType' object has no attribute 'message_types_by_name'报错", "file": "2023-04-24.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/169", "detail": "_HISTOGRAMPROTO = DESCRIPTOR.message_types_by_name['HistogramProto']", "id": 52} +{"title": "能否指定自己训练的text2vector模型?", "file": "2023-04-25.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/172", "detail": "请问大佬:", "id": 53} +{"title": "关于项目支持的模型以及quantization_bit潜在的影响的问题", "file": "2023-04-26.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/176", "detail": "作者您好~", "id": 54} +{"title": "运行python3.9 api.py WARNING: You must pass the application as an import string to enable 'reload' or 'workers'.", "file": "2023-04-26.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/179", "detail": "api.py文件最下面改成这样试试:", "id": 55} +{"title": "ValidationError: 1 validation error for HuggingFaceEmbeddings model_kwargs extra fields not permitted (type=value_error.extra)", "file": "2023-04-26.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/180", "detail": "ValidationError: 1 validation error for HuggingFaceEmbeddings", "id": 56} +{"title": "如果没有检索到相关性比较高的,回答“我不知道”", "file": "2023-04-26.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/181", "detail": "如果通过设计system_template,让模型在搜索到的文档都不太相关的情况下回答“我不知道”", "id": 57} +{"title": "请问如果不能联网,6B之类的文件从本地上传需要放到哪里", "file": "2023-04-26.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/182", "detail": "感谢大佬的项目,很有启发~", "id": 58} +{"title": "知识库问答--输入新的知识库名称是中文的话,会报error", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/184", "detail": "知识库问答--输入新的知识库名称是中文的话,会报error,选择要加载的知识库那里也不显示之前添加的知识库", "id": 59} +{"title": "现在能通过问题匹配的相似度值,来直接返回文档中的文段,而不经过模型吗?因为有些答案在文档中,模型自己回答,不能回答文档中的答案", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/186", "detail": "现在能通过问题匹配的相似度值,来直接返回文档中的文段,而不经过模型吗?因为有些答案在文档中,模型自己回答,不能回答文档中的答案。也就是说,提供向量检索回答+模型回答相结合的策略。如果相似度值高于一定数值,直接返回文档中的文本,没有高于就返回模型的回答或者不知道", "id": 60} +{"title": "TypeError: The type of ChatGLM.callback_manager differs from the new default value; if you wish to change the type of this field, please use a type annotation", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/188", "detail": "Mac 运行 python3 ./webui.py 报 TypeError: The type of ChatGLM.callback_manager differs from the new default value; if you wish to change the type of this field, please use a type annotation", "id": 61} +{"title": "Not Enough Memory", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/190", "detail": "运行命令行程序python cli_demo.py, 已经成功加载pdf文件, 报“DefaultCPUAllocator: not enough memory: you tried to allocate 458288380900 bytes”错误,请问哪里可以配置default memory", "id": 62} +{"title": "参与开发问题", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/191", "detail": "1.是否需要进专门的开发群", "id": 63} +{"title": "对话框中代码片段格式需改进", "file": "2023-04-27.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/192", "detail": "最好能改进下输出代码片段的格式,目前输出的格式还不友好。", "id": 64} +{"title": "请问未来有可能支持belle吗", "file": "2023-04-28.01", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/195", "detail": "如题,谢谢大佬", "id": 65} +{"title": "TypeError: cannot unpack non-iterable NoneType object", "file": "2023-04-28.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/200", "detail": "When i tried to change the knowledge vector store through `init_knowledge_vector_store`, the error `TypeError: cannot unpack non-iterable NoneType object` came out.", "id": 66} +{"title": "生成结果", "file": "2023-04-28.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/202", "detail": "你好,想问一下langchain+chatglm-6B,找到相似匹配的prompt,是直接返回prompt对应的答案信息,还是chatglm-6B在此基础上自己优化答案?", "id": 67} +{"title": "在win、ubuntu下都出现这个错误:attributeerror: 't5forconditionalgeneration' object has no attribute 'stream_chat'", "file": "2023-04-29.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/207", "detail": "在win、ubuntu。下载完模型后,没办法修改代码以执行本地模型,每次都要重新输入路径; LLM 模型、Embedding 模型支持也都在官网下的,在其他项目(wenda)下可以使用", "id": 68} +{"title": "[FEATURE] knowledge_based_chatglm.py: renamed or missing?", "file": "2023-04-30.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/210", "detail": "Not found. Was it renamed? Or, is it missing? How can I get it?", "id": 69} +{"title": "sudo apt-get install -y nvidia-container-toolkit-base执行报错", "file": "2023-05-01.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/211", "detail": "**问题描述 / Problem Description**", "id": 70} +{"title": "效果不佳几乎答不上来", "file": "2023-05-01.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/212", "detail": "提供了50条问答的docx文件", "id": 71} +{"title": "有没有可能新增一个基于chatglm api调用的方式构建langchain", "file": "2023-05-02.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/218", "detail": "我有两台8G GPU/40G内存的服务器,一个台做成了chatglm的api ;想基于另外一台服务器部署langchain;网上好像没有类似的代码。", "id": 72} +{"title": "电脑是intel的集成显卡; 运行时告知我找不到nvcuda.dll,模型无法运行", "file": "2023-05-02.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/219", "detail": "您好,我的电脑是intel的集成显卡,不过CPU是i5-11400 @ 2.60GHz ,内存64G;", "id": 73} +{"title": "根据langchain官方的文档和使用模式,是否可以改Faiss为Elasticsearch?会需要做哪些额外调整?求解", "file": "2023-05-03.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/221", "detail": "本人新手小白,由于业务模式的原因(有一些自己的场景和优化),希望利用Elasticsearch做这个体系内部的检索机制,不知道是否可以替换,同时,还会涉及到哪些地方的改动?或者说可能会有哪些其他影响,希望作者和大佬们不吝赐教!", "id": 74} +{"title": "请问未来有可能支持t5吗", "file": "2023-05-04.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/224", "detail": "请问可能支持基於t5的模型吗?", "id": 75} +{"title": "[BUG] 内存溢出 / torch.cuda.OutOfMemoryError:", "file": "2023-05-04.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/229", "detail": "**问题描述 / Problem Description**", "id": 76} +{"title": "报错 No module named 'chatglm_llm'", "file": "2023-05-04.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/230", "detail": "明明已经安装了包,却在python里吊不出来", "id": 77} +{"title": "能出一个api部署的描述文档吗", "file": "2023-05-04.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/233", "detail": "**功能描述 / Feature Description**", "id": 78} +{"title": "使用docs/API.md 出错", "file": "2023-05-04.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/234", "detail": "使用API.md文档2种方法,出错", "id": 79} +{"title": "加载pdf文档报错?", "file": "2023-05-05.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/238", "detail": "ew one with MEAN pooling.", "id": 80} +{"title": "上传的本地知识文件后再次上传不能显示,只显示成功了一个,别的上传成功后再次刷新就没了", "file": "2023-05-05.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/239", "detail": "您好,项目有很大启发,感谢~", "id": 81} +{"title": "创建了新的虚拟环境,安装了相关包,并且自动下载了相关的模型,但是仍旧出现:OSError: Unable to load weights from pytorch checkpoint file for", "file": "2023-05-05.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/240", "detail": "![78ac8e663fdc312d0e9d78da95925c4](https://user-images.githubusercontent.com/34124260/236378728-9ea4424f-0f7f-4013-9d33-820b723de321.png)", "id": 82} +{"title": "[BUG] 数据加载不进来", "file": "2023-05-05.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/243", "detail": "使用的.txt格式,utf-8编码,报以下错误", "id": 83} +{"title": "不能读取pdf", "file": "2023-05-05.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/244", "detail": "请问是webui还是cli_demo", "id": 84} +{"title": "本地txt文件有500M,加载的时候很慢,如何提高速度?", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/251", "detail": "![yayRzxSYHP](https://user-images.githubusercontent.com/109277248/236592902-f5ab338d-c1e9-43dc-ae16-9df2cd3c1378.jpg)", "id": 85} +{"title": "[BUG] gradio上传知识库后刷新之后 知识库就不见了 只有重启才能看到之前的上传的知识库", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/253", "detail": "gradio上传知识库后刷新之后 知识库就不见了 只有重启才能看到之前的上传的知识库", "id": 86} +{"title": "[FEATURE] 可以支持 OpenAI 的模型嘛?比如 GPT-3、GPT-3.5、GPT-4;embedding 增加 text-embedding-ada-002", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/254", "detail": "**功能描述 / Feature Description**", "id": 87} +{"title": "[FEATURE] 能否增加对于milvus向量数据库的支持 / Concise description of the feature", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/256", "detail": "**功能描述 / Feature Description**", "id": 88} +{"title": "CPU和GPU上跑,除了速度有区别,准确率效果回答上有区别吗?", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/259", "detail": "理论上没有区别", "id": 89} +{"title": "m1,请问在生成回答时怎么看是否使用了mps or cpu?", "file": "2023-05-06.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/260", "detail": "m1,请问在生成回答时怎么看是否使用了mps or cpu?", "id": 90} +{"title": "知识库一刷新就没了", "file": "2023-05-07.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/263", "detail": "知识库上传后刷新就没了", "id": 91} +{"title": "本地部署报没有模型", "file": "2023-05-07.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/267", "detail": "建议在下载llm和embedding模型至本地后在configs/model_config中写入模型本地存储路径后再运行", "id": 92} +{"title": "[BUG] python3: can't open file 'webui.py': [Errno 2] No such file or directory", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/269", "detail": "**问题描述 / Problem Description**", "id": 93} +{"title": "模块缺失提示", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/271", "detail": "因为已有自己使用的docker环境,直接启动webui.py,提示", "id": 94} +{"title": "运行api.py后,执行curl -X POST \"http://127.0.0.1:7861\" 报错?", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/272", "detail": "执行curl -X POST \"http://127.0.0.1:7861\" \\ -H 'Content-Type: application/json' \\ -d '{\"prompt\": \"你好\", \"history\": []}',报错怎么解决", "id": 95} +{"title": "[BUG] colab安装requirements提示protobuf版本问题?", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/273", "detail": "pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.", "id": 96} +{"title": "请问项目里面向量相似度使用了什么方法计算呀?", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/275", "detail": "基本按照langchain里的FAISS.similarity_search_with_score_by_vector实现", "id": 97} +{"title": "[BUG] 安装detectron2后,pdf无法加载", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/276", "detail": "**问题描述 / Problem Description**", "id": 98} +{"title": "[BUG] 使用ChatYuan-V2模型无法流式输出,会报错", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/277", "detail": "一方面好像是ChatYuan本身不支持stream_chat,有人在clueai那边提了issue他们说还没开发,所以估计这个attribute调不起来;但是另一方面看报错好像是T5模型本身就不是decoder-only模型,所以不能流式输出吧(个人理解)", "id": 99} +{"title": "[BUG] 无法加载text2vec模型", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/278", "detail": "**问题描述 / Problem Description**", "id": 100} +{"title": "请问能否增加网络搜索功能", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/281", "detail": "请问能否增加网络搜索功能", "id": 101} +{"title": "[FEATURE] 结构化数据sql、excel、csv啥时会支持呐。", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/283", "detail": "**功能描述 / Feature Description**", "id": 102} +{"title": "TypeError: ChatGLM._call() got an unexpected keyword argument 'stop'", "file": "2023-05-08.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/284", "detail": "No sentence-transformers model found with name D:\\DevProject\\langchain-ChatGLM\\GanymedeNil\\text2vec-large-chinese. Creating a new one with MEAN pooling.", "id": 103} +{"title": "关于api.py的一些bug和设计逻辑问题?", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/285", "detail": "首先冒昧的问一下,这个api.py,开发者大佬们是在自己电脑上测试后确实没问题吗?", "id": 104} +{"title": "有没有租用的算力平台上,运行api.py后,浏览器http://localhost:7861/报错", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/287", "detail": "是不是租用的gpu平台上都会出现这个问题???", "id": 105} +{"title": "请问一下项目中有用到文档段落切割方法吗?", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/288", "detail": "text_load中的文档切割方法用上了吗?在代码中看好像没有用到?", "id": 106} +{"title": "报错 raise ValueError(f\"Knowledge base {knowledge_base_id} not found\") ValueError: Knowledge base ./vector_store not found", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/289", "detail": "File \"/root/autodl-tmp/chatglm/langchain-ChatGLM-master/api.py\", line 183, in chat", "id": 107} +{"title": "能接入vicuna模型吗", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/290", "detail": "目前本地已经有了vicuna模型能直接接入吗?", "id": 108} +{"title": "[BUG] 提问公式相关问题大概率爆显存", "file": "2023-05-09.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/291", "detail": "**问题描述 / Problem Description**", "id": 109} +{"title": "安装pycocotools失败,找了好多方法都不能解决。", "file": "2023-05-10.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/292", "detail": "**问题描述 / Problem Description**", "id": 110} +{"title": "使用requirements安装,PyTorch安装的是CPU版本", "file": "2023-05-10.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/294", "detail": "如题目,使用requirements安装,PyTorch安装的是CPU版本,运行程序的时候,也是使用CPU在工作。", "id": 111} +{"title": "能不能给一个毛坯服务器的部署教程", "file": "2023-05-10.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/298", "detail": "“开发部署”你当成服务器的部署教程用就行了。", "id": 112} +{"title": " Error(s) in loading state_dict for ChatGLMForConditionalGeneration:", "file": "2023-05-10.02", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/299", "detail": "运行中出现的问题,7860的端口页面显示不出来,求助。", "id": 113} +{"title": "ChatYuan-large-v2模型加载失败", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/300", "detail": "**实际结果 / Actual Result**", "id": 114} +{"title": "新增摘要功能", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/303", "detail": "你好,后续会考虑新增对长文本信息进行推理和语音理解功能吗?比如生成摘要", "id": 115} +{"title": "[BUG] pip install -r requirements.txt 出错", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/304", "detail": "pip install langchain -i https://pypi.org/simple", "id": 116} +{"title": "[BUG] 上传知识库文件报错", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/305", "detail": "![19621e29eaa547d01213bee53d81e6a](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/84606552/7f6ceb46-e494-4b0e-939c-23b585a6d9d8)", "id": 117} +{"title": "[BUG] AssertionError: Component with id 41 not a valid input component.", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/306", "detail": "**问题描述 / Problem Description**", "id": 118} +{"title": "[BUG] CUDA out of memory with container deployment", "file": "2023-05-10.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/310", "detail": "**问题描述 / Problem Description**", "id": 119} +{"title": "[FEATURE] 增加微调训练功能", "file": "2023-05-11.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/311", "detail": "**功能描述 / Feature Description**", "id": 120} +{"title": "如何使用多卡部署,多个gpu", "file": "2023-05-11.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/315", "detail": "机器上有多个gpu,如何全使用了", "id": 121} +{"title": "请问这个知识库问答,和chatglm的关系是什么", "file": "2023-05-11.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/319", "detail": "这个知识库问答,哪部分关联到了chatglm,是不是没有这个chatglm,知识库问答也可单单拎出来", "id": 122} +{"title": "[BUG] 运行的时候报错ImportError: libcudnn.so.8: cannot open shared object file: No such file or directory", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/324", "detail": "**问题描述 / Problem Description**raceback (most recent call last):", "id": 123} +{"title": "webui启动成功,但有报错", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/325", "detail": "**问题描述 / Problem Description**", "id": 124} +{"title": "切换MOSS的时候报错", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/327", "detail": "danshi但是发布的源码中,", "id": 125} +{"title": "vicuna模型是否能接入?", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/328", "detail": "您好!关于MOSS模型和vicuna模型,都是AutoModelForCausalLM来加载模型的,但是稍作更改(模型路径这些)会报这个错误。这个错误的造成是什么", "id": 126} +{"title": "你好,请问一下在阿里云CPU服务器上跑可以吗?可以的话比较理想的cpu配置是什么?", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/330", "detail": "你好,请问一下在阿里云CPU服务器上跑可以吗?可以的话比较理想的cpu配置是什么?", "id": 127} +{"title": "你好,请问8核32g的CPU可以跑多轮对话吗?", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/331", "detail": "什么样的cpu配置比较好呢?我目前想部署CPU下的多轮对话?", "id": 128} +{"title": "[BUG] 聊天内容输入超过10000个字符系统出现错误", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/332", "detail": "聊天内容输入超过10000个字符系统出现错误,如下图所示:", "id": 129} +{"title": "能增加API的多用户访问接口部署吗?", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/333", "detail": "默认部署程序仅支持单用户访问,多用户则需要排队访问。测试过相关的几个Github多用户工程,但是其中一些仍然不满足要求。本节将系统介绍如何实现多用户同时访问ChatGLM的部署接口,包括http、websocket(流式输出,stream)和web页面等方式,主要目录如下所示。", "id": 130} +{"title": "多卡部署", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/334", "detail": "用单机多卡或多机多卡,fastapi部署模型,怎样提高并发", "id": 131} +{"title": "WEBUI能否指定知识库目录?", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/335", "detail": "**功能描述 / Feature Description**", "id": 132} +{"title": "[BUG] Cannot read properties of undefined (reading 'error')", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/336", "detail": "**问题描述 / Problem Description**", "id": 133} +{"title": "[BUG] 1 validation error for HuggingFaceEmbeddings model_kwargs extra fields not permitted.", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/337", "detail": "模型加载到 100% 后出现问题:", "id": 134} +{"title": "上传知识库需要重启能不能修复一下", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/338", "detail": "挺严重的这个问题", "id": 135} +{"title": "[BUG] 4块v100卡爆显存,在LLM会话模式也一样", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/339", "detail": "**问题描述 / Problem Description**", "id": 136} +{"title": "针对上传的文件配置不同的TextSpliter", "file": "2023-05-12.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/341", "detail": "1. 目前的ChineseTextSpliter切分对英文尤其是代码文件不友好,而且限制固定长度;导致对话结果不如人意", "id": 137} +{"title": "[FEATURE] 未来可增加Bloom系列模型吗?根据甲骨易的测试,这系列中文评测效果不错", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/346", "detail": "**功能描述 / Feature Description**", "id": 138} +{"title": "[BUG] v0.1.12打包镜像后启动webui.py失败 / Concise description of the issue", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/347", "detail": "**问题描述 / Problem Description**", "id": 139} +{"title": "切换MOSS模型时报错", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/349", "detail": "昨天问了下,说是transformers版本不对,需要4.30.0,发现没有这个版本,今天更新到4.29.1,依旧报错,错误如下", "id": 140} +{"title": "[BUG] pdf文档加载失败", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/350", "detail": "**问题描述 / Problem Description**", "id": 141} +{"title": "建议可以在后期增强一波注释,这样也有助于更多人跟进提PR", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/351", "detail": "知道作者和团队在疯狂更新审查代码,只是建议后续稳定后可以把核心代码进行一些注释的补充,从而能帮助更多人了解各个模块作者的思路从而提出更好的优化。", "id": 142} +{"title": "[FEATURE] MOSS 量化版支援", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/353", "detail": "**功能描述 / Feature Description**", "id": 143} +{"title": "[BUG] moss模型无法加载", "file": "2023-05-13.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/356", "detail": "**问题描述 / Problem Description**", "id": 144} +{"title": "[BUG] load_doc_qa.py 中的 load_file 函数有bug", "file": "2023-05-14.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/358", "detail": "原函数为:", "id": 145} +{"title": "[FEATURE] API模式,知识库加载优化", "file": "2023-05-14.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/359", "detail": "如题,当前版本,每次调用本地知识库接口,都将加载一次知识库,是否有更好的方式?", "id": 146} +{"title": "运行Python api.py脚本后端部署后,怎么使用curl命令调用?", "file": "2023-05-15.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/361", "detail": "也就是说,我现在想做个对话机器人,想和公司的前后端联调?怎么与前后端相互调用呢?可私信,有偿解答!!!", "id": 147} +{"title": "上传知识库需要重启能不能修复一下", "file": "2023-05-15.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/363", "detail": "上传知识库需要重启能不能修复一下", "id": 148} +{"title": "[BUG] pip install -r requirements.txt -i https://pypi.tuna.tsinghua.edu.cn/simple", "file": "2023-05-15.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/364", "detail": "我的python是3.8.5的", "id": 149} +{"title": "pip install gradio 报错", "file": "2023-05-15.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/367", "detail": "大佬帮我一下", "id": 150} +{"title": "[BUG] pip install gradio 一直卡不动", "file": "2023-05-15.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/369", "detail": "![aba82742dd9d4d242181662eb5027a7](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/84606552/cd9600d9-f6e7-46b7-b1be-30ed8b99f76b)", "id": 151} +{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/370", "detail": "初次加载本地知识库成功,但提问后,就无法重写加载本地知识库", "id": 152} +{"title": "[FEATURE] 简洁阐述功能 / Concise description of the feature", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/371", "detail": "**功能描述 / Feature Description**", "id": 153} +{"title": "在windows上,模型文件默认会安装到哪", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/372", "detail": "-------------------------------------------------------------------------------", "id": 154} +{"title": "[FEATURE] 兼顾对话管理", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/374", "detail": "如何在知识库检索的情况下,兼顾对话管理?", "id": 155} +{"title": "llm device: cpu embedding device: cpu", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/376", "detail": "**问题描述 / Problem Description**", "id": 156} +{"title": "[FEATURE] 简洁阐述功能 /文本文件的知识点之间使用什么分隔符可以分割?", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/377", "detail": "**功能描述 / Feature Description**", "id": 157} +{"title": "[BUG] 上传文件失败:PermissionError: [WinError 32] 另一个程序正在使用此文件,进程无法访问。", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/379", "detail": "**问题描述 / Problem Description**", "id": 158} +{"title": "[BUG] 执行python api.py 报错", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/383", "detail": "错误信息", "id": 159} +{"title": "model_kwargs extra fields not permitted (type=value_error.extra)", "file": "2023-05-16.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/384", "detail": "大家好,请问这个有遇到的么,?", "id": 160} +{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-05-17.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/385", "detail": "执行的时候出现了ls1 = [ls[0]]", "id": 161} +{"title": "[FEATURE] 性能优化", "file": "2023-05-17.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/388", "detail": "**功能描述 / Feature Description**", "id": 162} +{"title": "[BUG] Moss模型问答,RuntimeError: probability tensor contains either inf, nan or element < 0", "file": "2023-05-17.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/390", "detail": "**问题描述 / Problem Description**", "id": 163} +{"title": "有没有人知道v100GPU的32G显存,会报错吗?支持V100GPU吗?", "file": "2023-05-17.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/392", "detail": "**问题描述 / Problem Description**", "id": 164} +{"title": "针对于编码问题比如'gbk' codec can't encode character '\\xab' in position 14: illegal multibyte sequence粗浅的解决方法", "file": "2023-05-17.03", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/397", "detail": "**功能描述 / Feature Description**", "id": 165} +{"title": "Could not import sentence_transformers python package. Please install it with `pip install sentence_transformers`.", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/400", "detail": "**问题描述 / Problem Description**", "id": 166} +{"title": "支持模型问答与检索问答", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/401", "detail": "不同的query,根据意图不一致,回答也应该不一样。", "id": 167} +{"title": "文本分割的时候,能不能按照txt文件的每行进行分割,也就是按照换行符号\\n进行分割???", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/403", "detail": "下面的代码应该怎么修改?", "id": 168} +{"title": "local_doc_qa/local_doc_chat 接口响应是串行", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/405", "detail": "**问题描述 / Problem Description**", "id": 169} +{"title": "为什么找到出处了,但是还是无法回答该问题?", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/406", "detail": "![图片](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/3349611/1fc81d61-2409-4330-9065-fdda1a27c86a)", "id": 170} +{"title": "请问下:知识库测试中的:添加单条内容,如果换成文本导入是是怎样的格式?我发现添加单条内容测试效果很好.", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/412", "detail": "我发现在知识库测试中`添加单条内容`,并且勾选`禁止内容分句入库`,即使 `不开启上下文关联`的测试效果都非常好.", "id": 171} +{"title": "[BUG] 无法配置知识库", "file": "2023-05-18.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/413", "detail": "**问题描述 / Problem Description**", "id": 172} +{"title": "[BUG] 部署在阿里PAI平台的EAS上访问页面是白屏", "file": "2023-05-19.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/414", "detail": "**问题描述 / Problem Description**", "id": 173} +{"title": "API部署后调用/local_doc_qa/local_doc_chat 返回Knowledge base samples not found", "file": "2023-05-19.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/416", "detail": "入参", "id": 174} +{"title": "[FEATURE] 上传word另存为的txt文件报 'ascii' codec can't decode byte 0xb9 in position 6: ordinal not in range(128)", "file": "2023-05-20.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/421", "detail": "上传word另存为的txt文件报", "id": 175} +{"title": "创建保存的知识库刷新后没有出来,这个知识库是永久保存的吗?可以连外部的 向量知识库吗?", "file": "2023-05-21.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/422", "detail": "创建保存的知识库刷新后没有出来,这个知识库是永久保存的吗?可以连外部的 向量知识库吗?", "id": 176} +{"title": "[BUG] 用colab运行,无法加载模型,报错:'NoneType' object has no attribute 'message_types_by_name'", "file": "2023-05-21.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/423", "detail": "**问题描述 / Problem Description**", "id": 177} +{"title": "请问是否需要用到向量数据库?以及什么时候需要用到向量数据库?", "file": "2023-05-21.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/424", "detail": "目前用的是 text2vec , 请问是否需要用到向量数据库?以及什么时候需要用到向量数据库?", "id": 178} +{"title": "huggingface模型引用问题", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/427", "detail": "它最近似乎变成了一个Error?", "id": 179} +{"title": "你好,加载本地txt文件出现这个killed错误,TXT文件有100M左右大小。原因是?谢谢。", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/429", "detail": "\"929aca3b22b8cd74e997a87b61d241b\"", "id": 180} +{"title": "想请问一下,关于对本地知识的管理是如何管理?例如:通过http API接口添加数据 或者 删除某条数据", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/430", "detail": "例如:通过http API接口添加、删除、修改 某条数据。", "id": 181} +{"title": "[FEATURE] 双栏pdf识别问题", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/432", "detail": "试了一下模型,感觉对单栏pdf识别的准确性较高,但是由于使用的基本是ocr的技术,对一些双栏pdf论文识别出来有很多问题,请问有什么办法改善吗?", "id": 182} +{"title": "部署启动小问题,小弟初学求大佬解答", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/433", "detail": "1.python loader/image_loader.py时,提示ModuleNotFoundError: No module named 'configs',但是跑python webui.py还是还能跑", "id": 183} +{"title": "能否支持检测到目录下文档有增加而去增量加载文档,不影响前台对话,其实就是支持读写分离。如果能支持查询哪些文档向量化了,删除过时文档等就更好了,谢谢。", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/434", "detail": "**功能描述 / Feature Description**", "id": 184} +{"title": "[BUG] 简洁阐述问题 / windows 下cuda错误,请用https://github.com/Keith-Hon/bitsandbytes-windows.git", "file": "2023-05-22.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/435", "detail": "pip install git+https://github.com/Keith-Hon/bitsandbytes-windows.git", "id": 185} +{"title": "[BUG] from commit 33bbb47, Required library version not found: libbitsandbytes_cuda121_nocublaslt.so. Maybe you need to compile it from source?", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/438", "detail": "**问题描述 / Problem Description**", "id": 186} +{"title": "[BUG] 简洁阐述问题 / Concise description of the issue上传60m的txt文件报错,显示超时,请问这个能上传的文件大小有限制吗", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/439", "detail": "ERROR 2023-05-23 11:13:09,627-1d: Timeout reached while detecting encoding for ./docs/GLM模型格式数据.txt", "id": 187} +{"title": "[BUG] TypeError: issubclass() arg 1 must be a class", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/440", "detail": "**问题描述**", "id": 188} +{"title": "执行python3 webui.py后,一直提示”模型未成功加载,请到页面左上角\"模型配置\"选项卡中重新选择后点击\"加载模型\"按钮“", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/441", "detail": "**问题描述 / Problem Description**", "id": 189} +{"title": "是否能提供网页文档得导入支持", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/444", "detail": "现在很多都是在线文档作为协作得工具,所以通过URL导入在线文档需求更大", "id": 190} +{"title": "[BUG] history 索引问题", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/445", "detail": "在比较对话框的history和模型chat function 中的history时, 发现并不匹配,在传入 llm._call 时,history用的索引是不是有点问题,导致上一轮对话的内容并不输入给模型。", "id": 191} +{"title": "[BUG] moss_llm没有实现", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/447", "detail": "有些方法没支持,如history_len", "id": 192} +{"title": "请问langchain-ChatGLM如何删除一条本地知识库的数据?", "file": "2023-05-23.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/448", "detail": "例如:用户刚刚提交了一条错误的数据到本地知识库中了,现在如何在本地知识库从找到,并且对此删除。", "id": 193} +{"title": "[BUG] 简洁阐述问题 / UnboundLocalError: local variable 'resp' referenced before assignment", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/450", "detail": "在最新一版的代码中, 运行api.py 出现了以上错误(UnboundLocalError: local variable 'resp' referenced before assignment), 通过debug的方式观察到local_doc_qa.llm.generatorAnswer(prompt=question, history=history,streaming=True)可能不返回任何值。", "id": 194} +{"title": "请问有没有 PROMPT_TEMPLATE 能让模型不回答敏感问题", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/452", "detail": "## PROMPT_TEMPLATE问题", "id": 195} +{"title": "[BUG] 测试环境 Python 版本有误", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/456", "detail": "**问题描述 / Problem Description**", "id": 196} +{"title": "[BUG] webui 部署后样式不正确", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/458", "detail": "**问题描述 / Problem Description**", "id": 197} +{"title": "配置默认LLM模型的问题", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/459", "detail": "**问题描述 / Problem Description**", "id": 198} +{"title": "[FEATURE]是时候更新一下autoDL的镜像了", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/460", "detail": "如题,跑了下autoDL的镜像,发现是4.27号的,git pull新版本的代码功能+老的依赖环境,各种奇奇怪怪的问题。", "id": 199} +{"title": "[BUG] tag:0.1.13 以cpu模式下,想使用本地模型无法跑起来,各种路径参数问题", "file": "2023-05-24.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/462", "detail": "-------------------------------------------------------------------------------", "id": 200} +{"title": "[BUG] 有没有同学遇到过这个错!!!加载本地txt文件出现这个killed错误,TXT文件有100M左右大小。", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/463", "detail": "运行cli_demo.py。是本地的txt文件太大了吗?100M左右。", "id": 201} +{"title": "API版本能否提供WEBSOCKET的流式接口", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/464", "detail": "webui 版本中,采用了WS的流式输出,整体感知反应很快", "id": 202} +{"title": "[BUG] 安装bug记录", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/465", "detail": "按照[install文档](https://github.com/imClumsyPanda/langchain-ChatGLM/blob/master/docs/INSTALL.md)安装的,", "id": 203} +{"title": "VUE的pnmp i执行失败的修复-用npm i命令即可", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/466", "detail": "感谢作者!非常棒的应用,用的很开心。", "id": 204} +{"title": "请教个问题,有没有人知道cuda11.4是否支持???", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/467", "detail": "请教个问题,有没有人知道cuda11.4是否支持???", "id": 205} +{"title": "请问有实现多轮问答中基于问题的搜索上下文关联么", "file": "2023-05-25.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/468", "detail": "在基于知识库的多轮问答中,第一个问题讲述了一个主题,后续的问题描述没有包含这个主题的关键词,但又存在上下文的关联。如果用后续问题去搜索知识库有可能会搜索出无关的信息,从而导致大模型无法正确回答问题。请问这个项目要考虑这种情况吗?", "id": 206} +{"title": "[BUG] 内存不足的问题", "file": "2023-05-26.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/470", "detail": "我用了本地的chatglm-6b-int4模型,然后显示了内存不足(win10+32G内存+1080ti11G),一般需要多少内存才足够?这个bug应该如何解决?", "id": 207} +{"title": "[BUG] 纯内网环境安装pycocotools失败", "file": "2023-05-26.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/472", "detail": "**问题描述 / Problem Description**", "id": 208} +{"title": "[BUG] webui.py 重新加载模型会导致 KeyError", "file": "2023-05-26.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/473", "detail": "**问题描述 / Problem Description**", "id": 209} +{"title": "chatyuan无法使用", "file": "2023-05-26.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/475", "detail": "**问题描述 / Problem Description**", "id": 210} +{"title": "[BUG] 文本分割模型AliTextSplitter存在bug,会把“.”作为分割符", "file": "2023-05-26.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/476", "detail": "阿里达摩院的语义分割模型存在bug,默认会把\".”作为分割符进行分割而不管上下文语义。是否还有其他分割符则未知。建议的修改方案:把“.”统一替换为其他字符,分割后再替换回来。或者添加其他分割模型。", "id": 211} +{"title": "[BUG] RuntimeError: Error in faiss::FileIOReader::FileIOReader(const char*) a", "file": "2023-05-27.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/479", "detail": "**问题描述 / Problem Description**", "id": 212} +{"title": "[FEATURE] 安装,为什么conda create要额外指定路径 用-p ,而不是默认的/envs下面", "file": "2023-05-28.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/481", "detail": "##**功能描述 / Feature Description**", "id": 213} +{"title": "[小白求助] 通过Anaconda执行webui.py后,无法打开web链接", "file": "2023-05-28.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/485", "detail": "在执行webui.py命令后,http://0.0.0.0:7860复制到浏览器后无法打开,显示“无法访问此网站”。", "id": 214} +{"title": "[BUG] 使用 p-tuningv2后的模型,重新加载报错", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/486", "detail": "把p-tunningv2训练完后的相关文件放到了p-tunningv2文件夹下,勾选使用p-tuningv2点重新加载模型,控制台输错错误信息:", "id": 215} +{"title": "[小白求助] 服务器上执行webui.py后,在本地无法打开web链接", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/487", "detail": "此项目执行在xxx.xx.xxx.xxx服务器上,我在webui.py上的代码为 (demo", "id": 216} +{"title": "[FEATURE] 能不能支持VisualGLM-6B", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/488", "detail": "**功能描述 / Feature Description**", "id": 217} +{"title": "你好,问一下各位,后端api部署的时候,支持多用户同时问答吗???", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/489", "detail": "支持多用户的话,最多支持多少用户问答?根据硬件而定吧?", "id": 218} +{"title": "V100GPU显存占满,而利用率却为0,这是为什么?", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/491", "detail": "\"de45fe2b6cb76fa091b6e8f76a3de60\"", "id": 219} +{"title": "[求助] 如果在公司内部搭建产品知识库,使用INT-4模型,200人规模需要配置多少显存的服务器?", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/492", "detail": "如题,计划给公司搭一个在线知识库。", "id": 220} +{"title": "你好,请教个问题,目前问答回复需要20秒左右,如何提高速度?V10032G服务器。", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/493", "detail": "**问题描述 / Problem Description**", "id": 221} +{"title": "[FEATURE] 如何实现只匹配下文,而不要上文的结果", "file": "2023-05-29.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/494", "detail": "在构建自己的知识库时,主要采用问答对的形式,那么也就是我需要的回答是在我的问题下面的内容,但是目前设置了chunk_size的值以后匹配的是上下文的内容,但我实际并不需要上文的。为了实现更完整的展示下面的答案,我只能调大chunk_size的值,但实际上上文的一半内容都是我不需要的。也就是扔了一半没用的东西给prompt,在faiss.py中我也没找到这块的一些描述,请问该如何进行修改呢?", "id": 222} +{"title": "你好,问一下,我调用api.py部署,为什么用ip加端口可以使用postman调用,而改为域名使用postman无法调用?", "file": "2023-05-30.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/497", "detail": "![5ufBSWxLyF](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/109277248/70e2fbac-5699-48d0-b0d1-3dc84fd042c2)", "id": 223} +{"title": "调用api.py中的stream_chat,返回source_documents中出现中文乱码。", "file": "2023-05-30.04", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/498", "detail": "-------------------------------------------------------------------------------", "id": 224} +{"title": "[BUG] 捉个虫,api.py中的stream_chat解析json问题", "file": "2023-05-30.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/501", "detail": "**问题描述 / Problem Description**", "id": 225} +{"title": "windows本地部署遇到了omp错误", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/502", "detail": "**问题描述 / Problem Description**", "id": 226} +{"title": "[BUG] bug14 ,\"POST /local_doc_qa/upload_file HTTP/1.1\" 422 Unprocessable Entity", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/503", "detail": "上传的文件报错,返回错误,api.py", "id": 227} +{"title": "你好,请教个问题,api.py部署的时候,如何改为多线程调用?谢谢", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/505", "detail": "目前的api.py脚本不支持多线程", "id": 228} +{"title": "你好,请教一下。api.py部署的时候,能不能提供给后端流失返回结果。", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/507", "detail": "curl -X 'POST' \\", "id": 229} +{"title": "流式输出,流式接口,使用server-sent events技术。", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/508", "detail": "想这样一样,https://blog.csdn.net/weixin_43228814/article/details/130063010", "id": 230} +{"title": "计划增加流式输出功能吗?ChatGLM模型通过api方式调用响应时间慢怎么破,Fastapi流式接口来解惑,能快速提升响应速度", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/509", "detail": "**问题描述 / Problem Description**", "id": 231} +{"title": "[BUG] 知识库上传时发生ERROR (could not open xxx for reading: No such file or directory)", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/510", "detail": "**问题描述 / Problem Description**", "id": 232} +{"title": "api.py脚本打算增加SSE流式输出吗?", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/511", "detail": "curl调用的时候可以检测第一个字,从而提升回复的体验", "id": 233} +{"title": "[BUG] 使用tornado实现webSocket,可以多个客户端同时连接,并且实现流式回复,但是多个客户端同时使用,答案就很乱,是模型不支持多线程吗", "file": "2023-05-31.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/512", "detail": "import asyncio", "id": 234} +{"title": "支持 chinese_alpaca_plus_lora 吗 基于llama的", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/514", "detail": "支持 chinese_alpaca_plus_lora 吗 基于llama的,https://github.com/ymcui/Chinese-LLaMA-Alpaca这个项目的", "id": 235} +{"title": "[BUG] 现在能读图片的pdf了,但是文字的pdf反而读不了了,什么情况???", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/515", "detail": "**问题描述 / Problem Description**", "id": 236} +{"title": "在推理的过程中卡住不动,进程无法正常结束", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/516", "detail": "**问题描述 / Problem Description**", "id": 237} +{"title": "curl调用的时候,从第二轮开始,curl如何传参可以实现多轮对话?", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/517", "detail": "第一轮调用:", "id": 238} +{"title": "建议添加api.py部署后的日志管理功能?", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/518", "detail": "-------------------------------------------------------------------------------", "id": 239} +{"title": "有大佬知道,怎么多线程部署api.py脚本吗?", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/519", "detail": "api.py部署后,使用下面的请求,时间较慢,好像是单线程,如何改为多线程部署api.py:", "id": 240} +{"title": "[BUG] 上传文件到知识库 任何格式与内容都永远失败", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/520", "detail": "上传知识库的时候,传txt无法解析,就算是穿content/sample里的样例txt也无法解析,上传md、pdf等都无法加载,会持续性等待,等到了超过30分钟也不行。", "id": 241} +{"title": "关于prompt_template的问题", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/521", "detail": "请问这段prompt_template是什么意思,要怎么使用?可以给一个具体模板参考下吗?", "id": 242} +{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-06-01.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/522", "detail": "**问题描述 / Problem Description**", "id": 243} +{"title": "中文分词句号处理(关于表达金额之间的\".\")", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/523", "detail": "建议处理12.6亿元的这样的分词,最好别分成12 和6亿这样的,需要放到一起", "id": 244} +{"title": "ImportError: cannot import name 'inference' from 'paddle'", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/526", "detail": "在网上找了一圈,有说升级paddle的,我做了还是没有用,有说安装paddlepaddle的,我找了豆瓣的镜像源,但安装报错cannot detect archive format", "id": 245} +{"title": "[BUG] webscoket 接口串行问题(/local_doc_qa/stream-chat/{knowledge_base_id})", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/527", "detail": "**问题描述 / Problem Description**", "id": 246} +{"title": "[FEATURE] 刷新页面更新知识库列表", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/528", "detail": "**功能描述以及改进方案**", "id": 247} +{"title": "[BUG] 使用ptuning微调模型后,问答效果并不好", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/530", "detail": "### 未调用ptuning", "id": 248} +{"title": "[BUG] 多轮对话效果不佳", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/532", "detail": "在进行多轮对话的时候,无论设置的history_len是多少,效果都不好。事实上我将其设置成了最大值10,但在对话中,仍然无法实现多轮对话:", "id": 249} +{"title": "RuntimeError: MPS backend out of memory (MPS allocated: 18.00 GB, other allocations: 4.87 MB, max allowed: 18.13 GB)", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/533", "detail": "**问题描述**", "id": 250} +{"title": " 请大家重视这个issue!真正使用肯定是多用户并发问答,希望增加此功能!!!", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/534", "detail": "这得看你有多少显卡", "id": 251} +{"title": "在启动项目的时候如何使用到多张gpu啊?", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/535", "detail": "**在启动项目的时候如何使用到多张gpu啊?**", "id": 252} +{"title": " 使用流式输出的时候,curl调用的格式是什么?", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/536", "detail": "app.websocket(\"/local_doc_qa/stream-chat/{knowledge_base_id}\")(stream_chat)中的knowledge_base_id应该填什么???", "id": 253} +{"title": "使用本地 vicuna-7b模型启动错误", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/538", "detail": "环境: ubuntu 22.04 cuda 12.1 没有安装nccl,使用rtx2080与m60显卡并行计算", "id": 254} +{"title": "为什么会不调用GPU直接调用CPU呢", "file": "2023-06-02.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/539", "detail": "我的阿里云配置是16G显存,用默认代码跑webui.py时提示", "id": 255} +{"title": "上传多个文件时会互相覆盖", "file": "2023-06-03.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/541", "detail": "1、在同一个知识库中上传多个文件时会互相覆盖,无法结合多个文档的知识,有大佬知道怎么解决吗?", "id": 256} +{"title": "[BUG] ‘gcc’不是内部或外部命令/LLM对话只能持续一轮", "file": "2023-06-03.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/542", "detail": "No compiled kernel found.", "id": 257} +{"title": "以API模式启动项目却没有知识库的接口列表?", "file": "2023-06-04.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/544", "detail": "请问如何获取知识库的接口列表?如果没有需要自行编写的话,可不可以提供相关的获取方式,感谢", "id": 258} +{"title": "程序以API模式启动的时候,如何才能让接口以stream模式被调用呢?", "file": "2023-06-05.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/546", "detail": "作者您好,我在以API模式进行程序启动后,我发现接口响应时间很长,怎么样才能让接口以stream模式被调用呢?我想实现像webui模式的回答那样", "id": 259} +{"title": "关于原文中表格转为文本后数据相关度问题。", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/547", "detail": "原文中表格数据转换为文本,以 (X-Y:值;...) 的格式每一行组织成一句话,但这样做后发现相关度较低,效果很差,有何好的方案吗?", "id": 260} +{"title": "启动后LLM和知识库问答模式均只有最后一轮记录", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/548", "detail": "拉取最新代码,问答时,每次页面只显示最后一次问答记录,需要修改什么参数才可以保留历史记录?", "id": 261} +{"title": "提供system message配置,以便于让回答不要超出知识库范围", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/549", "detail": "**功能描述 / Feature Description**", "id": 262} +{"title": "[BUG] 使用p-tunningv2报错", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/551", "detail": "按照readme的指示把p-tunningv2训练完后的文件放到了p-tunningv2文件夹下,勾选使用p-tuningv2点重新加载模型,控制台提示错误信息:", "id": 263} +{"title": "[BUG] 智障,这么多问题,也好意思放出来,浪费时间", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/553", "detail": "。。。", "id": 264} +{"title": "[FEATURE] 我看代码文件中有一个ali_text_splitter.py,为什么不用他这个文本分割器了?", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/554", "detail": "我看代码文件中有一个ali_text_splitter.py,为什么不用他这个文本分割器了?", "id": 265} +{"title": "加载文档函数报错", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/557", "detail": "def load_file(filepath, sentence_size=SENTENCE_SIZE):", "id": 266} +{"title": "参考指引安装docker后,运行cli_demo.py,提示killed", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/558", "detail": "root@b3d1bd08095c:/chatGLM# python3 cli_demo.py", "id": 267} +{"title": "注意:如果安装错误,注意这两个包的版本 wandb==0.11.0 protobuf==3.18.3", "file": "2023-06-06.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/559", "detail": "Error1: 如果启动异常报错 `protobuf` 需要更新到 `protobuf==3.18.3 `", "id": 268} +{"title": "知识库对长文的知识相关度匹配不太理想有何优化方向", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/563", "detail": "我们可能录入一个文章有 1W 字,里面涉及这个文章主题的很多角度问题,我们针对他提问,他相关度匹配的内容和实际我们需要的答案相差很大怎么办。", "id": 269} +{"title": "使用stream-chat函数进行流式输出的时候,能使用curl调用吗?", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/565", "detail": "为什么下面这样调用会报错???", "id": 270} +{"title": "有大佬实践过 并行 或者 多线程 的部署方案吗?", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/566", "detail": "+1", "id": 271} +{"title": "多线程部署遇到问题?", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/567", "detail": "\"3d87bf74f0cf1a4820cc9e46b245859\"", "id": 272} +{"title": "[BUG] 用fastchat加载vicuna-13b模型进行知识库的问答有token的限制错误", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/569", "detail": "当我开启fastchat的vicuna-13b的api服务,然后config那里配置好(api本地测试过可以返回结果),然后知识库加载好之后(知识库大概有1000多个文档,用chatGLM可以正常推理),进行问答时出现token超过限制,就问了一句hello;", "id": 273} +{"title": "现在的添加知识库,文件多了总是报错,也不知道自己加载了哪些文件,报错后也不知道是全部失败还是一部分成功;希望能有个加载指定文件夹作为知识库的功能", "file": "2023-06-07.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/574", "detail": "**功能描述 / Feature Description**", "id": 274} +{"title": "[BUG] moss模型本地加载报错", "file": "2023-06-08.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/577", "detail": "moss模型本地加载报错:", "id": 275} +{"title": "加载本地moss模型报错Can't instantiate abstract class MOSSLLM with abstract methods _history_len", "file": "2023-06-08.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/578", "detail": "(vicuna) ps@ps[13:56:20]:/data/chat/langchain-ChatGLM2/langchain-ChatGLM-0.1.13$ python webui.py --model-dir local_models --model moss --no-remote-model", "id": 276} +{"title": "[FEATURE] 能增加在前端页面控制prompt_template吗?或是能支持前端页面选择使用哪个prompt?", "file": "2023-06-08.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/579", "detail": "目前只能在config里修改一个prompt,想在多个不同场景切换比较麻烦", "id": 277} +{"title": "[BUG] streamlit ui的bug,在增加知识库时会报错", "file": "2023-06-08.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/580", "detail": "**问题描述 / Problem Description**", "id": 278} +{"title": "[FEATURE] webui/webui_st可以支持history吗?目前仅能一次对话", "file": "2023-06-08.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/581", "detail": "试了下webui和webui_st都不支持历史对话啊,只能对话一次,不能默认开启所有history吗?", "id": 279} +{"title": "启动python cli_demo.py --model chatglm-6b-int4-qe报错", "file": "2023-06-09.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/585", "detail": "下载好模型,和相关依赖环境,之间运行`python cli_demo.py --model chatglm-6b-int4-qe`报错了:", "id": 280} +{"title": "重新构建知识库报错", "file": "2023-06-09.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/586", "detail": "**问题描述 / Problem Description**", "id": 281} +{"title": "[FEATURE] 能否屏蔽paddle,我不需要OCR,效果差依赖环境还很复杂", "file": "2023-06-09.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/587", "detail": "希望能不依赖paddle", "id": 282} +{"title": "question :文档向量化这个可以自己手动实现么?", "file": "2023-06-09.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/589", "detail": "现有公司级数据500G+,需要使用这个功能,请问如何手动实现这个向量化,然后并加载", "id": 283} +{"title": "view前端能进行流式的返回吗??", "file": "2023-06-09.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/590", "detail": "view前端能进行流式的返回吗??", "id": 284} +{"title": "[BUG] Load parallel cpu kernel failed, using default cpu kernel code", "file": "2023-06-11.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/594", "detail": "**问题描述 / Problem Description**", "id": 285} +{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-06-11.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/595", "detail": "**问题描述 / Problem Description**", "id": 286} +{"title": "我在上传本地知识库时提示KeyError: 'name'错误,本地知识库都是.txt文件,文件数量大约是2000+。", "file": "2023-06-12.05", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/597", "detail": "\"KError\"", "id": 287} +{"title": "model_config.py中有vicuna-13b-hf模型的配置信息,但是好像还是不可用?", "file": "2023-06-12.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/600", "detail": "@dongyihua543", "id": 288} +{"title": "ImportError: Using SOCKS proxy, but the 'socksio' package is not installed. Make sure to install httpx using `pip install httpx[socks]`.", "file": "2023-06-12.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/605", "detail": "应该代理问题,但是尝试了好多方法都解决不了,", "id": 289} +{"title": "[BUG] similarity_search_with_score_by_vector在找不到匹配的情况下出错", "file": "2023-06-12.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/607", "detail": "在设置匹配阈值 VECTOR_SEARCH_SCORE_THRESHOLD 的情况下,vectorstore会返回空,此时上述处理函数会出错", "id": 290} +{"title": "[FEATURE] 请问如何搭建英文知识库呢", "file": "2023-06-12.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/609", "detail": "**功能描述 / Feature Description**", "id": 291} +{"title": "谁有vicuna权重?llama转换之后的", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/611", "detail": "**问题描述 / Problem Description**", "id": 292} +{"title": "[FEATURE] API能实现上传文件夹的功能么?", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/612", "detail": "用户懒得全选所有的文件,就想上传个文件夹,请问下API能实现这个功能么?", "id": 293} +{"title": "请问在多卡部署后,上传单个文件作为知识库,用的是单卡在生成向量还是多卡?", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/614", "detail": "目前我检测我本地多卡部署的,好像生成知识库向量的时候用的还是单卡", "id": 294} +{"title": "[BUG] python webui.py提示非法指令", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/615", "detail": "(/data/conda-langchain [root@chatglm langchain-ChatGLM]# python webui.py", "id": 295} +{"title": "知识库文件跨行切分问题", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/616", "detail": "我的知识库文件txt文件,是一行一条知识,用\\n分行。", "id": 296} +{"title": "[FEATURE] bing搜索问答有流式的API么?", "file": "2023-06-13.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/617", "detail": "web端是有这个bing搜索回答,但api接口没有发现,大佬能给个提示么?", "id": 297} +{"title": "希望出一个macos m2的安装教程", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/620", "detail": "mac m2安装,模型加载成功了,知识库文件也上传成功了,但是一问答就会报错,报错内容如下", "id": 298} +{"title": "为【出处】提供高亮显示", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/621", "detail": "具体出处里面,对相关的内容高亮显示,不包含前后文。", "id": 299} +{"title": "[BUG] CPU运行cli_demo.py,不回答,hang住", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/622", "detail": "没有GPU;32G内存的ubuntu机器。", "id": 300} +{"title": "关于删除知识库里面的文档后,LLM知识库对话的时候还是会返回该被删除文档的内容", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/623", "detail": "如题,在vue前端成功执行删除知识库里面文档A.txt后,未能也在faiss索引中也删除该文档,LLM还是会返回这个A.txt的内容,并且以A.txt为出处,未能达到删除的效果", "id": 301} +{"title": "[BUG] 调用知识库进行问答,显存会一直叠加", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/625", "detail": "14G的显存,调用的chatglm-6b-int8模型,进行知识库问答时,最多问答四次就会爆显存了,观察了一下显存使用情况,每一次使用就会增加一次显存,请问这样是正常的吗?是否有什么配置需要开启可以解决这个问题?例如进行一次知识库问答清空上次问题的显存?", "id": 302} +{"title": "[BUG] web页面 重新构建数据库 失败,导致 原来的上传的数据库都没了", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/626", "detail": "web页面 重新构建数据库 失败,导致 原来的上传的数据库都没了", "id": 303} +{"title": "在CPU上运行webui.py报错Tensor on device cpu is not on the expected device meta!", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/627", "detail": "在CPU上运行python webui.py能启动,但最后有:RuntimeError: Tensor on device cpu is not on the expected device meta!", "id": 304} +{"title": "OSError: [WinError 1114] 动态链接库(DLL)初始化例程失败。 Error loading \"E:\\xxx\\envs\\langchain\\lib\\site-packages\\torch\\lib\\caffe2_nvrtc.dll\" or one of its dependencies.哪位大佬知道如何解决吗?", "file": "2023-06-14.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/629", "detail": "**问题描述 / Problem Description**", "id": 305} +{"title": "[BUG] WEBUI删除知识库文档,会导致知识库问答失败", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/632", "detail": "如题,从知识库已有文件中选择要删除的文件,点击删除后,在问答框输入内容回车报错", "id": 306} +{"title": "更新后的版本中,删除知识库中的文件,再提问出现error错误", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/634", "detail": "针对更新版本,识别到一个问题,过程如下:", "id": 307} +{"title": "我配置好了环境,想要实现本地知识库的问答?可是它返回给我的", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/637", "detail": "没有总结,只有相关度的回复,但是我看演示里面表现的,回复是可以实现总结的,我去查询代码", "id": 308} +{"title": "[BUG] NPM run dev can not successfully start the VUE frontend", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/638", "detail": "**问题描述 / Problem Description**", "id": 309} +{"title": "[BUG] 简洁阐述问题 / Concise description of the issue", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/639", "detail": "**问题描述 / Problem Description**", "id": 310} +{"title": "提一个模型加载的bug,我在截图中修复了,你们有空可以看一下。", "file": "2023-06-15.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/642", "detail": "![model_load_bug](https://github.com/imClumsyPanda/langchain-ChatGLM/assets/59411575/4432adc4-ccdd-45d9-aafc-5f2d1963403b)", "id": 311} +{"title": "[求助]关于设置embedding model路径的问题", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/643", "detail": "如题,我之前成功跑起来过一次,但因环境丢失重新配置 再运行webui就总是报错", "id": 312} +{"title": "Lora微调后的模型可以直接使用吗", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/646", "detail": "看model_config.py里是有USE_LORA这个参数的,但是在cli_demo.py和webui.py这两个里面都没有用到,实际测试下来模型没有微调的效果,想问问现在这个功能实现了吗", "id": 313} +{"title": "write_check_file在tmp_files目录下生成的load_file.txt是否需要一直保留,占用空间很大,在建完索引后能否删除", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/647", "detail": "**功能描述 / Feature Description**", "id": 314} +{"title": "[BUG] /local_doc_qa/list_files?knowledge_base_id=test删除知识库bug", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/649", "detail": "1.新建test知识库并上传文件(在vue前端完成并检查后端发现确实生成了test文件夹以及下面的content和vec_store", "id": 315} +{"title": "[BUG] vue webui无法加载知识库", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/650", "detail": "拉取了最新的代码,分别运行了后端api和前端web,点击知识库,始终只能显示simple,无法加载知识库", "id": 316} +{"title": "不能本地加载moss模型吗?", "file": "2023-06-16.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/652", "detail": "手动下载模型设置local_model_path路径依旧提示缺少文件,该如何正确配置?", "id": 317} +{"title": "macos m2 pro docker 安装失败", "file": "2023-06-17.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/654", "detail": "macos m2 pro docker 安装失败", "id": 318} +{"title": " [BUG] mac m1 pro 运行提示 zsh: segmentation fault", "file": "2023-06-17.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/655", "detail": "运行: python webui.py", "id": 319} +{"title": "安装 requirements 报错", "file": "2023-06-17.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/656", "detail": "(langchainchatglm) D:\\github\\langchain-ChatGLM>pip install -r requirements.txt -i https://pypi.tuna.tsinghua.edu.cn/simple/", "id": 320} +{"title": "[BUG] AssertionError", "file": "2023-06-17.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/658", "detail": "**问题描述 / Problem Description**", "id": 321} +{"title": "[FEATURE] 支持AMD win10 本地部署吗?", "file": "2023-06-18.06", "url": "https://github.com/imClumsyPanda/langchain-ChatGLM/issues/660", "detail": "**功能描述 / Feature Description**", "id": 322} diff --git a/requirements.txt b/requirements.txt index 0af4bd18..c44ddf88 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,6 +1,6 @@ -langchain>=0.0.324 +langchain>=0.0.326 langchain-experimental>=0.0.30 -fschat[model_worker]==0.2.31 +fschat[model_worker]==0.2.32 xformers>=0.0.22.post4 openai>=0.28.1 sentence_transformers @@ -48,7 +48,7 @@ pandas~=2.0.3 streamlit>=1.26.0 streamlit-option-menu>=0.3.6 streamlit-antd-components>=0.1.11 -streamlit-chatbox==1.1.11 +streamlit-chatbox>=1.1.11 streamlit-aggrid>=0.3.4.post3 httpx[brotli,http2,socks]>=0.25.0 watchdog diff --git a/requirements_api.txt b/requirements_api.txt index 126e65f5..dbd110b3 100644 --- a/requirements_api.txt +++ b/requirements_api.txt @@ -1,10 +1,10 @@ -langchain>=0.0.324 +langchain>=0.0.326 langchain-experimental>=0.0.30 -fschat[model_worker]==0.2.31 +fschat[model_worker]==0.2.32 xformers>=0.0.22.post4 openai>=0.28.1 sentence_transformers>=2.2.2 -transformers>=4.34 +transformers>=4.35 torch>=2.1 torchvision torchaudio diff --git a/requirements_lite.txt b/requirements_lite.txt index 8b0a32d1..ebc35520 100644 --- a/requirements_lite.txt +++ b/requirements_lite.txt @@ -1,12 +1,12 @@ -langchain>=0.0.324 -fschat>=0.2.31 +langchain>=0.0.326 +fschat>=0.2.32 openai # sentence_transformers -# transformers>=4.33.0 +# transformers>=4.35.0 # torch>=2.0.1 # torchvision # torchaudio -fastapi>=0.103.1 +fastapi>=0.104.1 python-multipart nltk~=3.8.1 uvicorn~=0.23.1 diff --git a/server/agent/tools/search_all_knowledge_more.py b/server/agent/tools/search_all_knowledge_more.py index 7da9081f..a42681a7 100644 --- a/server/agent/tools/search_all_knowledge_more.py +++ b/server/agent/tools/search_all_knowledge_more.py @@ -71,15 +71,16 @@ bigdata,大数据的就业情况如何 这些数据库是你能访问的,冒号之前是他们的名字,冒号之后是他们的功能,你应该参考他们的功能来帮助你思考 + {database_names} 你的回答格式应该按照下面的内容,请注意```text 等标记都必须输出,这是我用来提取答案的标记。 - +不要输出中文的逗号,不要输出引号。 Question: ${{用户的问题}} ```text -${{知识库名称,查询问题,不要带有任何除了,之外的符号}} +${{知识库名称,查询问题,不要带有任何除了,之外的符号,比如不要输出中文的逗号,不要输出引号}} ```output 数据库查询的结果 @@ -165,7 +166,11 @@ class LLMKnowledgeChain(LLMChain): lines = cleaned_input_str.split("\n") # 使用逗号分割每一行,然后形成一个(数据库,查询)元组的列表 - queries = [(line.split(",")[0].strip(), line.split(",")[1].strip()) for line in lines] + try: + queries = [(line.split(",")[0].strip(), line.split(",")[1].strip()) for line in lines] + except: + queries = [(line.split(",")[0].strip(), line.split(",")[1].strip()) for line in lines] + print(queries) run_manager.on_text("知识库查询询内容:\n\n" + str(queries) + " \n\n", color="blue", verbose=self.verbose) output = self._evaluate_expression(queries) run_manager.on_text("\nAnswer: ", verbose=self.verbose) @@ -193,7 +198,10 @@ class LLMKnowledgeChain(LLMChain): cleaned_input_str = ( expression.replace("\"", "").replace("“", "").replace("”", "").replace("```", "").strip()) lines = cleaned_input_str.split("\n") - queries = [(line.split(",")[0].strip(), line.split(",")[1].strip()) for line in lines] + try: + queries = [(line.split(",")[0].strip(), line.split(",")[1].strip()) for line in lines] + except: + queries = [(line.split(",")[0].strip(), line.split(",")[1].strip()) for line in lines] await run_manager.on_text("知识库查询询内容:\n\n" + str(queries) + " \n\n", color="blue", verbose=self.verbose) diff --git a/server/agent/tools_select.py b/server/agent/tools_select.py index e0be597a..330db43d 100644 --- a/server/agent/tools_select.py +++ b/server/agent/tools_select.py @@ -20,21 +20,25 @@ tools = [ func=weathercheck, name="天气查询工具", description="无需访问互联网,使用这个工具查询中国各地未来24小时的天气", + args_schema=WhetherSchema, ), Tool.from_function( func=shell, name="shell工具", description="使用命令行工具输出", + args_schema=ShellInput, ), Tool.from_function( func=knowledge_search_more, name="知识库查询工具", description="优先访问知识库来获取答案", + args_schema=KnowledgeSearchInput, ), Tool.from_function( func=search_internet, name="互联网查询工具", description="如果你无法访问互联网,这个工具可以帮助你访问Bing互联网来解答问题", + args_schema=SearchInternetInput, ), Tool.from_function( func=wolfram, diff --git a/server/chat/agent_chat.py b/server/chat/agent_chat.py index 70478503..5c5983c0 100644 --- a/server/chat/agent_chat.py +++ b/server/chat/agent_chat.py @@ -76,7 +76,7 @@ async def agent_chat(query: str = Body(..., description="用户输入", examples agent = LLMSingleActionAgent( llm_chain=llm_chain, output_parser=output_parser, - stop=["\nObservation:", "Observation:", "<|im_end|>", "<|observation|>"], + stop=["\nObservation:", "Observation:", "<|im_end|>", "<|observation|>"], # Qwen模型中使用这个 allowed_tools=tool_names, ) # 把history转成agent的memory diff --git a/server/chat/chat.py b/server/chat/chat.py index 72255768..8659af1b 100644 --- a/server/chat/chat.py +++ b/server/chat/chat.py @@ -11,7 +11,7 @@ from langchain.prompts.chat import ChatPromptTemplate from typing import List, Optional from server.chat.utils import History from server.utils import get_prompt_template -from server.db.repository.chat_history_repository import add_chat_history_to_db, update_chat_history +from server.db.repository import add_chat_history_to_db, update_chat_history async def chat(query: str = Body(..., description="用户输入", examples=["恼羞成怒"]), @@ -81,4 +81,4 @@ async def chat(query: str = Body(..., description="用户输入", examples=["恼 history=history, model_name=model_name, prompt_name=prompt_name), - media_type="text/event-stream") + media_type="text/event-stream") \ No newline at end of file diff --git a/server/chat/knowledge_base_chat.py b/server/chat/knowledge_base_chat.py index cffb3d30..664217bf 100644 --- a/server/chat/knowledge_base_chat.py +++ b/server/chat/knowledge_base_chat.py @@ -105,4 +105,4 @@ async def knowledge_base_chat(query: str = Body(..., description="用户输入", history=history, model_name=model_name, prompt_name=prompt_name), - media_type="text/event-stream") + media_type="text/event-stream") \ No newline at end of file diff --git a/server/db/repository/__init__.py b/server/db/repository/__init__.py index e69de29b..aad591ea 100644 --- a/server/db/repository/__init__.py +++ b/server/db/repository/__init__.py @@ -0,0 +1,3 @@ +from .chat_history_repository import * +from .knowledge_base_repository import * +from .knowledge_file_repository import * \ No newline at end of file diff --git a/server/knowledge_base/kb_doc_api.py b/server/knowledge_base/kb_doc_api.py index 11b8f45b..abad0ccf 100644 --- a/server/knowledge_base/kb_doc_api.py +++ b/server/knowledge_base/kb_doc_api.py @@ -368,7 +368,7 @@ def recreate_vector_store( "code": 200, "msg": f"({i + 1} / {len(files)}): {file_name}", "total": len(files), - "finished": i, + "finished": i + 1, "doc": file_name, }, ensure_ascii=False) kb.add_doc(kb_file, not_refresh_vs_cache=True) diff --git a/server/utils.py b/server/utils.py index c31bdc18..99fbac11 100644 --- a/server/utils.py +++ b/server/utils.py @@ -15,7 +15,6 @@ import httpx from typing import Literal, Optional, Callable, Generator, Dict, Any, Awaitable, Union - async def wrap_done(fn: Awaitable, event: asyncio.Event): """Wrap an awaitable with a event to signal when it's done or an exception is raised.""" try: @@ -41,61 +40,21 @@ def get_ChatOpenAI( ) -> ChatOpenAI: ## 以下模型是Langchain原生支持的模型,这些模型不会走Fschat封装 config_models = list_config_llm_models() - if model_name in config_models.get("langchain", {}): - config = config_models["langchain"][model_name] - if model_name == "Azure-OpenAI": - model = AzureChatOpenAI( - streaming=streaming, - verbose=verbose, - callbacks=callbacks, - deployment_name=config.get("deployment_name"), - model_version=config.get("model_version"), - openai_api_type=config.get("openai_api_type"), - openai_api_base=config.get("api_base_url"), - openai_api_version=config.get("api_version"), - openai_api_key=config.get("api_key"), - openai_proxy=config.get("openai_proxy"), - temperature=temperature, - max_tokens=max_tokens, - ) - elif model_name == "OpenAI": - model = ChatOpenAI( - streaming=streaming, - verbose=verbose, - callbacks=callbacks, - model_name=config.get("model_name"), - openai_api_base=config.get("api_base_url"), - openai_api_key=config.get("api_key"), - openai_proxy=config.get("openai_proxy"), - temperature=temperature, - max_tokens=max_tokens, - ) - elif model_name == "Anthropic": - model = ChatAnthropic( - streaming=streaming, - verbose=verbose, - callbacks=callbacks, - model_name=config.get("model_name"), - anthropic_api_key=config.get("api_key"), - - ) - ## TODO 支持其他的Langchain原生支持的模型 - else: - ## 非Langchain原生支持的模型,走Fschat封装 - config = get_model_worker_config(model_name) - model = ChatOpenAI( - streaming=streaming, - verbose=verbose, - callbacks=callbacks, - openai_api_key=config.get("api_key", "EMPTY"), - openai_api_base=config.get("api_base_url", fschat_openai_api_address()), - model_name=model_name, - temperature=temperature, - max_tokens=max_tokens, - openai_proxy=config.get("openai_proxy"), - **kwargs - ) + ## 非Langchain原生支持的模型,走Fschat封装 + config = get_model_worker_config(model_name) + model = ChatOpenAI( + streaming=streaming, + verbose=verbose, + callbacks=callbacks, + openai_api_key=config.get("api_key", "EMPTY"), + openai_api_base=config.get("api_base_url", fschat_openai_api_address()), + model_name=model_name, + temperature=temperature, + max_tokens=max_tokens, + openai_proxy=config.get("openai_proxy"), + **kwargs + ) return model @@ -249,9 +208,9 @@ def torch_gc(): empty_cache() except Exception as e: msg = ("如果您使用的是 macOS 建议将 pytorch 版本升级至 2.0.0 或更高版本," - "以支持及时清理 torch 产生的内存占用。") + "以支持及时清理 torch 产生的内存占用。") logger.error(f'{e.__class__.__name__}: {msg}', - exc_info=e if log_verbose else None) + exc_info=e if log_verbose else None) except Exception: ... @@ -437,7 +396,6 @@ def get_model_worker_config(model_name: str = None) -> dict: config.update(ONLINE_LLM_MODEL.get(model_name, {}).copy()) config.update(FSCHAT_MODEL_WORKERS.get(model_name, {}).copy()) - if model_name in ONLINE_LLM_MODEL: config["online_api"] = True if provider := config.get("provider"): @@ -474,7 +432,7 @@ def fschat_controller_address() -> str: def fschat_model_worker_address(model_name: str = LLM_MODEL) -> str: - if model := get_model_worker_config(model_name): # TODO: depends fastchat + if model := get_model_worker_config(model_name): # TODO: depends fastchat host = model["host"] if host == "0.0.0.0": host = "127.0.0.1" @@ -624,7 +582,7 @@ def run_in_thread_pool( thread = pool.submit(func, **kwargs) tasks.append(thread) - for obj in as_completed(tasks): # TODO: Ctrl+c无法停止 + for obj in as_completed(tasks): # TODO: Ctrl+c无法停止 yield obj.result() diff --git a/webui_pages/dialogue/dialogue.py b/webui_pages/dialogue/dialogue.py index 823b8fba..80b0d945 100644 --- a/webui_pages/dialogue/dialogue.py +++ b/webui_pages/dialogue/dialogue.py @@ -328,4 +328,4 @@ def dialogue_page(api: ApiRequest, is_lite: bool = False): file_name=f"{now:%Y-%m-%d %H.%M}_对话记录.md", mime="text/markdown", use_container_width=True, - ) + ) \ No newline at end of file