diff --git a/LICENSE b/LICENSE
new file mode 100644
index 0000000..0ad25db
--- /dev/null
+++ b/LICENSE
@@ -0,0 +1,661 @@
+ GNU AFFERO GENERAL PUBLIC LICENSE
+ Version 3, 19 November 2007
+
+ Copyright (C) 2007 Free Software Foundation, Inc.
+ Everyone is permitted to copy and distribute verbatim copies
+ of this license document, but changing it is not allowed.
+
+ Preamble
+
+ The GNU Affero General Public License is a free, copyleft license for
+software and other kinds of works, specifically designed to ensure
+cooperation with the community in the case of network server software.
+
+ The licenses for most software and other practical works are designed
+to take away your freedom to share and change the works. By contrast,
+our General Public Licenses are intended to guarantee your freedom to
+share and change all versions of a program--to make sure it remains free
+software for all its users.
+
+ When we speak of free software, we are referring to freedom, not
+price. Our General Public Licenses are designed to make sure that you
+have the freedom to distribute copies of free software (and charge for
+them if you wish), that you receive source code or can get it if you
+want it, that you can change the software or use pieces of it in new
+free programs, and that you know you can do these things.
+
+ Developers that use our General Public Licenses protect your rights
+with two steps: (1) assert copyright on the software, and (2) offer
+you this License which gives you legal permission to copy, distribute
+and/or modify the software.
+
+ A secondary benefit of defending all users' freedom is that
+improvements made in alternate versions of the program, if they
+receive widespread use, become available for other developers to
+incorporate. Many developers of free software are heartened and
+encouraged by the resulting cooperation. However, in the case of
+software used on network servers, this result may fail to come about.
+The GNU General Public License permits making a modified version and
+letting the public access it on a server without ever releasing its
+source code to the public.
+
+ The GNU Affero General Public License is designed specifically to
+ensure that, in such cases, the modified source code becomes available
+to the community. It requires the operator of a network server to
+provide the source code of the modified version running there to the
+users of that server. Therefore, public use of a modified version, on
+a publicly accessible server, gives the public access to the source
+code of the modified version.
+
+ An older license, called the Affero General Public License and
+published by Affero, was designed to accomplish similar goals. This is
+a different license, not a version of the Affero GPL, but Affero has
+released a new version of the Affero GPL which permits relicensing under
+this license.
+
+ The precise terms and conditions for copying, distribution and
+modification follow.
+
+ TERMS AND CONDITIONS
+
+ 0. Definitions.
+
+ "This License" refers to version 3 of the GNU Affero General Public License.
+
+ "Copyright" also means copyright-like laws that apply to other kinds of
+works, such as semiconductor masks.
+
+ "The Program" refers to any copyrightable work licensed under this
+License. Each licensee is addressed as "you". "Licensees" and
+"recipients" may be individuals or organizations.
+
+ To "modify" a work means to copy from or adapt all or part of the work
+in a fashion requiring copyright permission, other than the making of an
+exact copy. The resulting work is called a "modified version" of the
+earlier work or a work "based on" the earlier work.
+
+ A "covered work" means either the unmodified Program or a work based
+on the Program.
+
+ To "propagate" a work means to do anything with it that, without
+permission, would make you directly or secondarily liable for
+infringement under applicable copyright law, except executing it on a
+computer or modifying a private copy. Propagation includes copying,
+distribution (with or without modification), making available to the
+public, and in some countries other activities as well.
+
+ To "convey" a work means any kind of propagation that enables other
+parties to make or receive copies. Mere interaction with a user through
+a computer network, with no transfer of a copy, is not conveying.
+
+ An interactive user interface displays "Appropriate Legal Notices"
+to the extent that it includes a convenient and prominently visible
+feature that (1) displays an appropriate copyright notice, and (2)
+tells the user that there is no warranty for the work (except to the
+extent that warranties are provided), that licensees may convey the
+work under this License, and how to view a copy of this License. If
+the interface presents a list of user commands or options, such as a
+menu, a prominent item in the list meets this criterion.
+
+ 1. Source Code.
+
+ The "source code" for a work means the preferred form of the work
+for making modifications to it. "Object code" means any non-source
+form of a work.
+
+ A "Standard Interface" means an interface that either is an official
+standard defined by a recognized standards body, or, in the case of
+interfaces specified for a particular programming language, one that
+is widely used among developers working in that language.
+
+ The "System Libraries" of an executable work include anything, other
+than the work as a whole, that (a) is included in the normal form of
+packaging a Major Component, but which is not part of that Major
+Component, and (b) serves only to enable use of the work with that
+Major Component, or to implement a Standard Interface for which an
+implementation is available to the public in source code form. A
+"Major Component", in this context, means a major essential component
+(kernel, window system, and so on) of the specific operating system
+(if any) on which the executable work runs, or a compiler used to
+produce the work, or an object code interpreter used to run it.
+
+ The "Corresponding Source" for a work in object code form means all
+the source code needed to generate, install, and (for an executable
+work) run the object code and to modify the work, including scripts to
+control those activities. However, it does not include the work's
+System Libraries, or general-purpose tools or generally available free
+programs which are used unmodified in performing those activities but
+which are not part of the work. For example, Corresponding Source
+includes interface definition files associated with source files for
+the work, and the source code for shared libraries and dynamically
+linked subprograms that the work is specifically designed to require,
+such as by intimate data communication or control flow between those
+subprograms and other parts of the work.
+
+ The Corresponding Source need not include anything that users
+can regenerate automatically from other parts of the Corresponding
+Source.
+
+ The Corresponding Source for a work in source code form is that
+same work.
+
+ 2. Basic Permissions.
+
+ All rights granted under this License are granted for the term of
+copyright on the Program, and are irrevocable provided the stated
+conditions are met. This License explicitly affirms your unlimited
+permission to run the unmodified Program. The output from running a
+covered work is covered by this License only if the output, given its
+content, constitutes a covered work. This License acknowledges your
+rights of fair use or other equivalent, as provided by copyright law.
+
+ You may make, run and propagate covered works that you do not
+convey, without conditions so long as your license otherwise remains
+in force. You may convey covered works to others for the sole purpose
+of having them make modifications exclusively for you, or provide you
+with facilities for running those works, provided that you comply with
+the terms of this License in conveying all material for which you do
+not control copyright. Those thus making or running the covered works
+for you must do so exclusively on your behalf, under your direction
+and control, on terms that prohibit them from making any copies of
+your copyrighted material outside their relationship with you.
+
+ Conveying under any other circumstances is permitted solely under
+the conditions stated below. Sublicensing is not allowed; section 10
+makes it unnecessary.
+
+ 3. Protecting Users' Legal Rights From Anti-Circumvention Law.
+
+ No covered work shall be deemed part of an effective technological
+measure under any applicable law fulfilling obligations under article
+11 of the WIPO copyright treaty adopted on 20 December 1996, or
+similar laws prohibiting or restricting circumvention of such
+measures.
+
+ When you convey a covered work, you waive any legal power to forbid
+circumvention of technological measures to the extent such circumvention
+is effected by exercising rights under this License with respect to
+the covered work, and you disclaim any intention to limit operation or
+modification of the work as a means of enforcing, against the work's
+users, your or third parties' legal rights to forbid circumvention of
+technological measures.
+
+ 4. Conveying Verbatim Copies.
+
+ You may convey verbatim copies of the Program's source code as you
+receive it, in any medium, provided that you conspicuously and
+appropriately publish on each copy an appropriate copyright notice;
+keep intact all notices stating that this License and any
+non-permissive terms added in accord with section 7 apply to the code;
+keep intact all notices of the absence of any warranty; and give all
+recipients a copy of this License along with the Program.
+
+ You may charge any price or no price for each copy that you convey,
+and you may offer support or warranty protection for a fee.
+
+ 5. Conveying Modified Source Versions.
+
+ You may convey a work based on the Program, or the modifications to
+produce it from the Program, in the form of source code under the
+terms of section 4, provided that you also meet all of these conditions:
+
+ a) The work must carry prominent notices stating that you modified
+ it, and giving a relevant date.
+
+ b) The work must carry prominent notices stating that it is
+ released under this License and any conditions added under section
+ 7. This requirement modifies the requirement in section 4 to
+ "keep intact all notices".
+
+ c) You must license the entire work, as a whole, under this
+ License to anyone who comes into possession of a copy. This
+ License will therefore apply, along with any applicable section 7
+ additional terms, to the whole of the work, and all its parts,
+ regardless of how they are packaged. This License gives no
+ permission to license the work in any other way, but it does not
+ invalidate such permission if you have separately received it.
+
+ d) If the work has interactive user interfaces, each must display
+ Appropriate Legal Notices; however, if the Program has interactive
+ interfaces that do not display Appropriate Legal Notices, your
+ work need not make them do so.
+
+ A compilation of a covered work with other separate and independent
+works, which are not by their nature extensions of the covered work,
+and which are not combined with it such as to form a larger program,
+in or on a volume of a storage or distribution medium, is called an
+"aggregate" if the compilation and its resulting copyright are not
+used to limit the access or legal rights of the compilation's users
+beyond what the individual works permit. Inclusion of a covered work
+in an aggregate does not cause this License to apply to the other
+parts of the aggregate.
+
+ 6. Conveying Non-Source Forms.
+
+ You may convey a covered work in object code form under the terms
+of sections 4 and 5, provided that you also convey the
+machine-readable Corresponding Source under the terms of this License,
+in one of these ways:
+
+ a) Convey the object code in, or embodied in, a physical product
+ (including a physical distribution medium), accompanied by the
+ Corresponding Source fixed on a durable physical medium
+ customarily used for software interchange.
+
+ b) Convey the object code in, or embodied in, a physical product
+ (including a physical distribution medium), accompanied by a
+ written offer, valid for at least three years and valid for as
+ long as you offer spare parts or customer support for that product
+ model, to give anyone who possesses the object code either (1) a
+ copy of the Corresponding Source for all the software in the
+ product that is covered by this License, on a durable physical
+ medium customarily used for software interchange, for a price no
+ more than your reasonable cost of physically performing this
+ conveying of source, or (2) access to copy the
+ Corresponding Source from a network server at no charge.
+
+ c) Convey individual copies of the object code with a copy of the
+ written offer to provide the Corresponding Source. This
+ alternative is allowed only occasionally and noncommercially, and
+ only if you received the object code with such an offer, in accord
+ with subsection 6b.
+
+ d) Convey the object code by offering access from a designated
+ place (gratis or for a charge), and offer equivalent access to the
+ Corresponding Source in the same way through the same place at no
+ further charge. You need not require recipients to copy the
+ Corresponding Source along with the object code. If the place to
+ copy the object code is a network server, the Corresponding Source
+ may be on a different server (operated by you or a third party)
+ that supports equivalent copying facilities, provided you maintain
+ clear directions next to the object code saying where to find the
+ Corresponding Source. Regardless of what server hosts the
+ Corresponding Source, you remain obligated to ensure that it is
+ available for as long as needed to satisfy these requirements.
+
+ e) Convey the object code using peer-to-peer transmission, provided
+ you inform other peers where the object code and Corresponding
+ Source of the work are being offered to the general public at no
+ charge under subsection 6d.
+
+ A separable portion of the object code, whose source code is excluded
+from the Corresponding Source as a System Library, need not be
+included in conveying the object code work.
+
+ A "User Product" is either (1) a "consumer product", which means any
+tangible personal property which is normally used for personal, family,
+or household purposes, or (2) anything designed or sold for incorporation
+into a dwelling. In determining whether a product is a consumer product,
+doubtful cases shall be resolved in favor of coverage. For a particular
+product received by a particular user, "normally used" refers to a
+typical or common use of that class of product, regardless of the status
+of the particular user or of the way in which the particular user
+actually uses, or expects or is expected to use, the product. A product
+is a consumer product regardless of whether the product has substantial
+commercial, industrial or non-consumer uses, unless such uses represent
+the only significant mode of use of the product.
+
+ "Installation Information" for a User Product means any methods,
+procedures, authorization keys, or other information required to install
+and execute modified versions of a covered work in that User Product from
+a modified version of its Corresponding Source. The information must
+suffice to ensure that the continued functioning of the modified object
+code is in no case prevented or interfered with solely because
+modification has been made.
+
+ If you convey an object code work under this section in, or with, or
+specifically for use in, a User Product, and the conveying occurs as
+part of a transaction in which the right of possession and use of the
+User Product is transferred to the recipient in perpetuity or for a
+fixed term (regardless of how the transaction is characterized), the
+Corresponding Source conveyed under this section must be accompanied
+by the Installation Information. But this requirement does not apply
+if neither you nor any third party retains the ability to install
+modified object code on the User Product (for example, the work has
+been installed in ROM).
+
+ The requirement to provide Installation Information does not include a
+requirement to continue to provide support service, warranty, or updates
+for a work that has been modified or installed by the recipient, or for
+the User Product in which it has been modified or installed. Access to a
+network may be denied when the modification itself materially and
+adversely affects the operation of the network or violates the rules and
+protocols for communication across the network.
+
+ Corresponding Source conveyed, and Installation Information provided,
+in accord with this section must be in a format that is publicly
+documented (and with an implementation available to the public in
+source code form), and must require no special password or key for
+unpacking, reading or copying.
+
+ 7. Additional Terms.
+
+ "Additional permissions" are terms that supplement the terms of this
+License by making exceptions from one or more of its conditions.
+Additional permissions that are applicable to the entire Program shall
+be treated as though they were included in this License, to the extent
+that they are valid under applicable law. If additional permissions
+apply only to part of the Program, that part may be used separately
+under those permissions, but the entire Program remains governed by
+this License without regard to the additional permissions.
+
+ When you convey a copy of a covered work, you may at your option
+remove any additional permissions from that copy, or from any part of
+it. (Additional permissions may be written to require their own
+removal in certain cases when you modify the work.) You may place
+additional permissions on material, added by you to a covered work,
+for which you have or can give appropriate copyright permission.
+
+ Notwithstanding any other provision of this License, for material you
+add to a covered work, you may (if authorized by the copyright holders of
+that material) supplement the terms of this License with terms:
+
+ a) Disclaiming warranty or limiting liability differently from the
+ terms of sections 15 and 16 of this License; or
+
+ b) Requiring preservation of specified reasonable legal notices or
+ author attributions in that material or in the Appropriate Legal
+ Notices displayed by works containing it; or
+
+ c) Prohibiting misrepresentation of the origin of that material, or
+ requiring that modified versions of such material be marked in
+ reasonable ways as different from the original version; or
+
+ d) Limiting the use for publicity purposes of names of licensors or
+ authors of the material; or
+
+ e) Declining to grant rights under trademark law for use of some
+ trade names, trademarks, or service marks; or
+
+ f) Requiring indemnification of licensors and authors of that
+ material by anyone who conveys the material (or modified versions of
+ it) with contractual assumptions of liability to the recipient, for
+ any liability that these contractual assumptions directly impose on
+ those licensors and authors.
+
+ All other non-permissive additional terms are considered "further
+restrictions" within the meaning of section 10. If the Program as you
+received it, or any part of it, contains a notice stating that it is
+governed by this License along with a term that is a further
+restriction, you may remove that term. If a license document contains
+a further restriction but permits relicensing or conveying under this
+License, you may add to a covered work material governed by the terms
+of that license document, provided that the further restriction does
+not survive such relicensing or conveying.
+
+ If you add terms to a covered work in accord with this section, you
+must place, in the relevant source files, a statement of the
+additional terms that apply to those files, or a notice indicating
+where to find the applicable terms.
+
+ Additional terms, permissive or non-permissive, may be stated in the
+form of a separately written license, or stated as exceptions;
+the above requirements apply either way.
+
+ 8. Termination.
+
+ You may not propagate or modify a covered work except as expressly
+provided under this License. Any attempt otherwise to propagate or
+modify it is void, and will automatically terminate your rights under
+this License (including any patent licenses granted under the third
+paragraph of section 11).
+
+ However, if you cease all violation of this License, then your
+license from a particular copyright holder is reinstated (a)
+provisionally, unless and until the copyright holder explicitly and
+finally terminates your license, and (b) permanently, if the copyright
+holder fails to notify you of the violation by some reasonable means
+prior to 60 days after the cessation.
+
+ Moreover, your license from a particular copyright holder is
+reinstated permanently if the copyright holder notifies you of the
+violation by some reasonable means, this is the first time you have
+received notice of violation of this License (for any work) from that
+copyright holder, and you cure the violation prior to 30 days after
+your receipt of the notice.
+
+ Termination of your rights under this section does not terminate the
+licenses of parties who have received copies or rights from you under
+this License. If your rights have been terminated and not permanently
+reinstated, you do not qualify to receive new licenses for the same
+material under section 10.
+
+ 9. Acceptance Not Required for Having Copies.
+
+ You are not required to accept this License in order to receive or
+run a copy of the Program. Ancillary propagation of a covered work
+occurring solely as a consequence of using peer-to-peer transmission
+to receive a copy likewise does not require acceptance. However,
+nothing other than this License grants you permission to propagate or
+modify any covered work. These actions infringe copyright if you do
+not accept this License. Therefore, by modifying or propagating a
+covered work, you indicate your acceptance of this License to do so.
+
+ 10. Automatic Licensing of Downstream Recipients.
+
+ Each time you convey a covered work, the recipient automatically
+receives a license from the original licensors, to run, modify and
+propagate that work, subject to this License. You are not responsible
+for enforcing compliance by third parties with this License.
+
+ An "entity transaction" is a transaction transferring control of an
+organization, or substantially all assets of one, or subdividing an
+organization, or merging organizations. If propagation of a covered
+work results from an entity transaction, each party to that
+transaction who receives a copy of the work also receives whatever
+licenses to the work the party's predecessor in interest had or could
+give under the previous paragraph, plus a right to possession of the
+Corresponding Source of the work from the predecessor in interest, if
+the predecessor has it or can get it with reasonable efforts.
+
+ You may not impose any further restrictions on the exercise of the
+rights granted or affirmed under this License. For example, you may
+not impose a license fee, royalty, or other charge for exercise of
+rights granted under this License, and you may not initiate litigation
+(including a cross-claim or counterclaim in a lawsuit) alleging that
+any patent claim is infringed by making, using, selling, offering for
+sale, or importing the Program or any portion of it.
+
+ 11. Patents.
+
+ A "contributor" is a copyright holder who authorizes use under this
+License of the Program or a work on which the Program is based. The
+work thus licensed is called the contributor's "contributor version".
+
+ A contributor's "essential patent claims" are all patent claims
+owned or controlled by the contributor, whether already acquired or
+hereafter acquired, that would be infringed by some manner, permitted
+by this License, of making, using, or selling its contributor version,
+but do not include claims that would be infringed only as a
+consequence of further modification of the contributor version. For
+purposes of this definition, "control" includes the right to grant
+patent sublicenses in a manner consistent with the requirements of
+this License.
+
+ Each contributor grants you a non-exclusive, worldwide, royalty-free
+patent license under the contributor's essential patent claims, to
+make, use, sell, offer for sale, import and otherwise run, modify and
+propagate the contents of its contributor version.
+
+ In the following three paragraphs, a "patent license" is any express
+agreement or commitment, however denominated, not to enforce a patent
+(such as an express permission to practice a patent or covenant not to
+sue for patent infringement). To "grant" such a patent license to a
+party means to make such an agreement or commitment not to enforce a
+patent against the party.
+
+ If you convey a covered work, knowingly relying on a patent license,
+and the Corresponding Source of the work is not available for anyone
+to copy, free of charge and under the terms of this License, through a
+publicly available network server or other readily accessible means,
+then you must either (1) cause the Corresponding Source to be so
+available, or (2) arrange to deprive yourself of the benefit of the
+patent license for this particular work, or (3) arrange, in a manner
+consistent with the requirements of this License, to extend the patent
+license to downstream recipients. "Knowingly relying" means you have
+actual knowledge that, but for the patent license, your conveying the
+covered work in a country, or your recipient's use of the covered work
+in a country, would infringe one or more identifiable patents in that
+country that you have reason to believe are valid.
+
+ If, pursuant to or in connection with a single transaction or
+arrangement, you convey, or propagate by procuring conveyance of, a
+covered work, and grant a patent license to some of the parties
+receiving the covered work authorizing them to use, propagate, modify
+or convey a specific copy of the covered work, then the patent license
+you grant is automatically extended to all recipients of the covered
+work and works based on it.
+
+ A patent license is "discriminatory" if it does not include within
+the scope of its coverage, prohibits the exercise of, or is
+conditioned on the non-exercise of one or more of the rights that are
+specifically granted under this License. You may not convey a covered
+work if you are a party to an arrangement with a third party that is
+in the business of distributing software, under which you make payment
+to the third party based on the extent of your activity of conveying
+the work, and under which the third party grants, to any of the
+parties who would receive the covered work from you, a discriminatory
+patent license (a) in connection with copies of the covered work
+conveyed by you (or copies made from those copies), or (b) primarily
+for and in connection with specific products or compilations that
+contain the covered work, unless you entered into that arrangement,
+or that patent license was granted, prior to 28 March 2007.
+
+ Nothing in this License shall be construed as excluding or limiting
+any implied license or other defenses to infringement that may
+otherwise be available to you under applicable patent law.
+
+ 12. No Surrender of Others' Freedom.
+
+ If conditions are imposed on you (whether by court order, agreement or
+otherwise) that contradict the conditions of this License, they do not
+excuse you from the conditions of this License. If you cannot convey a
+covered work so as to satisfy simultaneously your obligations under this
+License and any other pertinent obligations, then as a consequence you may
+not convey it at all. For example, if you agree to terms that obligate you
+to collect a royalty for further conveying from those to whom you convey
+the Program, the only way you could satisfy both those terms and this
+License would be to refrain entirely from conveying the Program.
+
+ 13. Remote Network Interaction; Use with the GNU General Public License.
+
+ Notwithstanding any other provision of this License, if you modify the
+Program, your modified version must prominently offer all users
+interacting with it remotely through a computer network (if your version
+supports such interaction) an opportunity to receive the Corresponding
+Source of your version by providing access to the Corresponding Source
+from a network server at no charge, through some standard or customary
+means of facilitating copying of software. This Corresponding Source
+shall include the Corresponding Source for any work covered by version 3
+of the GNU General Public License that is incorporated pursuant to the
+following paragraph.
+
+ Notwithstanding any other provision of this License, you have
+permission to link or combine any covered work with a work licensed
+under version 3 of the GNU General Public License into a single
+combined work, and to convey the resulting work. The terms of this
+License will continue to apply to the part which is the covered work,
+but the work with which it is combined will remain governed by version
+3 of the GNU General Public License.
+
+ 14. Revised Versions of this License.
+
+ The Free Software Foundation may publish revised and/or new versions of
+the GNU Affero General Public License from time to time. Such new versions
+will be similar in spirit to the present version, but may differ in detail to
+address new problems or concerns.
+
+ Each version is given a distinguishing version number. If the
+Program specifies that a certain numbered version of the GNU Affero General
+Public License "or any later version" applies to it, you have the
+option of following the terms and conditions either of that numbered
+version or of any later version published by the Free Software
+Foundation. If the Program does not specify a version number of the
+GNU Affero General Public License, you may choose any version ever published
+by the Free Software Foundation.
+
+ If the Program specifies that a proxy can decide which future
+versions of the GNU Affero General Public License can be used, that proxy's
+public statement of acceptance of a version permanently authorizes you
+to choose that version for the Program.
+
+ Later license versions may give you additional or different
+permissions. However, no additional obligations are imposed on any
+author or copyright holder as a result of your choosing to follow a
+later version.
+
+ 15. Disclaimer of Warranty.
+
+ THERE IS NO WARRANTY FOR THE PROGRAM, TO THE EXTENT PERMITTED BY
+APPLICABLE LAW. EXCEPT WHEN OTHERWISE STATED IN WRITING THE COPYRIGHT
+HOLDERS AND/OR OTHER PARTIES PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY
+OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING, BUT NOT LIMITED TO,
+THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
+PURPOSE. THE ENTIRE RISK AS TO THE QUALITY AND PERFORMANCE OF THE PROGRAM
+IS WITH YOU. SHOULD THE PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF
+ALL NECESSARY SERVICING, REPAIR OR CORRECTION.
+
+ 16. Limitation of Liability.
+
+ IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING
+WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MODIFIES AND/OR CONVEYS
+THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, INCLUDING ANY
+GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING OUT OF THE
+USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED TO LOSS OF
+DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY YOU OR THIRD
+PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER PROGRAMS),
+EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE POSSIBILITY OF
+SUCH DAMAGES.
+
+ 17. Interpretation of Sections 15 and 16.
+
+ If the disclaimer of warranty and limitation of liability provided
+above cannot be given local legal effect according to their terms,
+reviewing courts shall apply local law that most closely approximates
+an absolute waiver of all civil liability in connection with the
+Program, unless a warranty or assumption of liability accompanies a
+copy of the Program in return for a fee.
+
+ END OF TERMS AND CONDITIONS
+
+ How to Apply These Terms to Your New Programs
+
+ If you develop a new program, and you want it to be of the greatest
+possible use to the public, the best way to achieve this is to make it
+free software which everyone can redistribute and change under these terms.
+
+ To do so, attach the following notices to the program. It is safest
+to attach them to the start of each source file to most effectively
+state the exclusion of warranty; and each file should have at least
+the "copyright" line and a pointer to where the full notice is found.
+
+
+ Copyright (C)
+
+ This program is free software: you can redistribute it and/or modify
+ it under the terms of the GNU Affero General Public License as published
+ by the Free Software Foundation, either version 3 of the License, or
+ (at your option) any later version.
+
+ This program is distributed in the hope that it will be useful,
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
+ GNU Affero General Public License for more details.
+
+ You should have received a copy of the GNU Affero General Public License
+ along with this program. If not, see .
+
+Also add information on how to contact you by electronic and paper mail.
+
+ If your software can interact with users remotely through a computer
+network, you should also make sure that it provides a way for users to
+get its source. For example, if your program is a web application, its
+interface could display a "Source" link that leads users to an archive
+of the code. There are many ways you could offer source, and different
+solutions will be better for different programs; see section 13 for the
+specific requirements.
+
+ You should also get your employer (if you work as a programmer) or school,
+if any, to sign a "copyright disclaimer" for the program, if necessary.
+For more information on this, and how to apply and follow the GNU AGPL, see
+.
diff --git a/llm_quantize/quantize.py b/llm_quantize/quantize.py
index 7fb1c24..708b6c8 100644
--- a/llm_quantize/quantize.py
+++ b/llm_quantize/quantize.py
@@ -1,6 +1,6 @@
+import os
import subprocess
import sys
-import os
def execute_commands(model_dir_path, quantization=None):
@@ -13,7 +13,7 @@ def execute_commands(model_dir_path, quantization=None):
if quantization:
model_file = f"llama.cpp/models/{model_dir_path}/ggml-model-f16.gguf"
quantized_model_file = f"llama.cpp/models/{model_dir_path.split('/')[-1]}/ggml-model-{quantization}.gguf"
- subprocess.run(["llama.cpp/llm_quantize", model_file, quantized_model_file, quantization], check=True)
+ subprocess.run(["llama.cpp/quantize", model_file, quantized_model_file, quantization], check=True)
else:
print("llama.cpp doesn't exist, check readme how to clone.")
diff --git a/projects/Basic-RAG/BasciRAG_CustomPrompt.py b/projects/Basic-RAG/BasciRAG_CustomPrompt.py
index 4222aad..3a3eed9 100644
--- a/projects/Basic-RAG/BasciRAG_CustomPrompt.py
+++ b/projects/Basic-RAG/BasciRAG_CustomPrompt.py
@@ -1,5 +1,5 @@
+from grag.components.prompt import Prompt
from grag.rag.basic_rag import BasicRAG
-from grap.components.prompt import Prompt
custom_prompt = Prompt(
input_keys={"context", "question"},
diff --git a/projects/Basic-RAG/BasicRAG-ingest_data.py b/projects/Basic-RAG/BasicRAG-ingest_data.py
deleted file mode 100644
index 5d2417b..0000000
--- a/projects/Basic-RAG/BasicRAG-ingest_data.py
+++ /dev/null
@@ -1,87 +0,0 @@
-import os
-from pathlib import Path
-from uuid import UUID, uuid5
-
-from grag.components.multivec_retriever import Retriever
-from grag.components.parse_pdf import ParsePDF
-from grag.components.utils import get_config
-from tqdm import tqdm
-
-config = get_config()
-
-DRY_RUN = False
-
-data_path = Path(config['data']['data_path']) / 'pdf'
-formats_to_add = ['Text', 'Tables']
-glob_pattern = '**/*.pdf'
-
-namespace = UUID('8c9040b0-b5cd-4d7c-bc2e-737da1b24ebf')
-record_filename = uuid5(namespace, str(data_path)) # Unique file record name based on folder path
-
-records_dir = Path(config['data']['data_path']) / 'records'
-records_dir.mkdir(parents=True, exist_ok=True)
-record_file = records_dir / f'{record_filename}.txt'
-
-
-def load_processed_files():
- # Load processed files from a file if it exists
- if os.path.exists(record_file):
- with open(record_file, 'r') as file_record:
- processed_files.update(file_record.read().splitlines())
-
-
-def update_processed_file_record(file_path, dry_run=False):
- # Update (append) the processed file record file
- with open(record_file, 'a') as file:
- if not dry_run:
- file.write(file_path + '\n')
-
-
-def add_file_to_database(file_path: Path, dry_run=False):
- # Check if file_path is in the processed file set
- if str(file_path) not in processed_files:
- # Add file to the vector database
- add_to_database(file_path, dry_run=dry_run)
- # Add file_path to the processed file set
- processed_files.add(str(file_path))
- # Update the processed file record file
- update_processed_file_record(str(file_path), dry_run=dry_run)
- return f'Completed adding - {file_path.relative_to(data_path)}'
- else:
- return f'Already exists - {file_path.relative_to(data_path)}'
-
-
-def add_to_database(file_path, dry_run=False):
- if not dry_run:
- docs = parser.load_file(file_path)
- for format_key in formats_to_add:
- retriever.add_docs(docs[format_key])
-
-
-parser = ParsePDF()
-retriever = Retriever()
-
-processed_files = set()
-load_processed_files() # Load processed files into the set on script startup
-
-
-def main():
- filepath_gen = data_path.glob(glob_pattern)
- num_files = len(list(data_path.glob(glob_pattern)))
- print(f'DATA PATH : {data_path}')
- print(f'No of PDFs to add: {num_files}')
- pbar = tqdm(filepath_gen, total=num_files, desc='Adding Files ')
- for file in pbar:
- pbar.set_postfix({'Current file': file.relative_to(data_path)})
- pbar.write(add_file_to_database(file, dry_run=DRY_RUN))
- # if str(file) not in processed_files:
- # add_to_database(file, dry_run=DRY_RUN) # Add file to the vector database
- # processed_files.add(str(file)) # Add file_path to processed set
- # update_processed_file_record(str(file), dry_run=DRY_RUN) # Update the processed file record file
- # pbar.write(f'Completed adding - {file.relative_to(data_path)}')
- # else:
- # pbar.write(f'Already exists - {file.relative_to(data_path)}')
-
-
-if __name__ == "__main__":
- main()
diff --git a/projects/Basic-RAG/BasicRAG_FewShotPrompt.py b/projects/Basic-RAG/BasicRAG_FewShotPrompt.py
index cd9199c..5fc2d46 100644
--- a/projects/Basic-RAG/BasicRAG_FewShotPrompt.py
+++ b/projects/Basic-RAG/BasicRAG_FewShotPrompt.py
@@ -1,5 +1,5 @@
+from grag.components.prompt import FewShotPrompt
from grag.rag.basic_rag import BasicRAG
-from grap.components.prompt import FewShotPrompt
custom_few_shot_prompt = FewShotPrompt(
input_keys={"context", "question"},
diff --git a/projects/Basic-RAG/BasicRAG_ingest.py b/projects/Basic-RAG/BasicRAG_ingest.py
new file mode 100644
index 0000000..b80e6ef
--- /dev/null
+++ b/projects/Basic-RAG/BasicRAG_ingest.py
@@ -0,0 +1,14 @@
+from pathlib import Path
+
+from grag.components.multivec_retriever import Retriever
+from grag.components.vectordb.deeplake_client import DeepLakeClient
+
+# from grag.rag.basic_rag import BasicRAG
+
+client = DeepLakeClient(collection_name="test")
+retriever = Retriever(vectordb=client)
+
+dir_path = Path(__file__).parent / 'some_dir'
+
+retriever.ingest(dir_path)
+# rag = BasicRAG(doc_chain="refine")
diff --git a/projects/Basic-RAG/BasicRAG_refine.py b/projects/Basic-RAG/BasicRAG_refine.py
index cce6398..ad8d8e7 100644
--- a/projects/Basic-RAG/BasicRAG_refine.py
+++ b/projects/Basic-RAG/BasicRAG_refine.py
@@ -1,5 +1,9 @@
-from grag.grag.rag import BasicRAG
+from grag.components.multivec_retriever import Retriever
+from grag.components.vectordb.deeplake_client import DeepLakeClient
+from grag.rag.basic_rag import BasicRAG
+client = DeepLakeClient(collection_name="test")
+retriever = Retriever(vectordb=client)
rag = BasicRAG(doc_chain="refine")
if __name__ == "__main__":
diff --git a/projects/Basic-RAG/BasicRAG_stuff.py b/projects/Basic-RAG/BasicRAG_stuff.py
index 4bfafc3..63edeab 100644
--- a/projects/Basic-RAG/BasicRAG_stuff.py
+++ b/projects/Basic-RAG/BasicRAG_stuff.py
@@ -1,6 +1,10 @@
-from grag.grag.rag import BasicRAG
+from grag.components.multivec_retriever import Retriever
+from grag.components.vectordb.deeplake_client import DeepLakeClient
+from grag.rag.basic_rag import BasicRAG
-rag = BasicRAG(doc_chain="stuff")
+client = DeepLakeClient(collection_name="test")
+retriever = Retriever(vectordb=client)
+rag = BasicRAG(doc_chain="stuff", retriever=retriever)
if __name__ == "__main__":
while True:
diff --git a/projects/Basic-RAG/BasicRAG_v1_depricated.py b/projects/Basic-RAG/BasicRAG_v1_depricated.py
deleted file mode 100644
index 63b5b89..0000000
--- a/projects/Basic-RAG/BasicRAG_v1_depricated.py
+++ /dev/null
@@ -1,42 +0,0 @@
-from pathlib import Path
-
-from src.components.llm import LLM, llm_conf
-from src.components.multivec_retriever import Retriever
-from src.components.utils import stuff_docs, load_prompt
-
-# from prompts import
-'''
-Basic RAG v1 - stuff, chunks
- Given a query, retrieve similar chunks from vector database. Concat them into a single string, called context.
- Using the prompt template, call llm. Return chunk sources.
-'''
-
-llm_ = LLM()
-# llm = llm_.load_model(pipeline='hf', is_local=False)
-llm = llm_.load_model()
-
-retriever = Retriever(top_k=3)
-
-prompt_name = 'Llama-2_QA_1.json'
-prompt_path = Path(__file__).parent / 'prompts' / prompt_name
-prompt_template = load_prompt(prompt_path)
-
-
-def call_rag(query):
- retrieved_docs = retriever.get_chunk(query)
- context = stuff_docs(retrieved_docs)
- prompt = prompt_template.format(context=context, question=query)
- response = llm.invoke(prompt)
- sources = [doc.metadata["source"] for doc in retrieved_docs]
- return response, sources
-
-
-if __name__ == "__main__":
- while True:
- query = input("Query:")
- response, sources = call_rag(query)
- if not llm_conf['std_out']:
- print(response)
- print(f'Sources: ')
- for index, source in enumerate(sources):
- print(f'\t{index}: {source}')
diff --git a/projects/Basic-RAG/BasicRAG_v2_depricated.py b/projects/Basic-RAG/BasicRAG_v2_depricated.py
deleted file mode 100644
index fd2e250..0000000
--- a/projects/Basic-RAG/BasicRAG_v2_depricated.py
+++ /dev/null
@@ -1,52 +0,0 @@
-from pathlib import Path
-
-from src.components.llm import LLM, llm_conf
-from src.components.multivec_retriever import Retriever
-from src.components.utils import load_prompt
-
-
-'''
-Basic RAG v2 - refine, top_k
-
-'''
-
-llm_ = LLM()
-llm = llm_.load_model()
-
-retriever = Retriever(top_k=3)
-
-prompts_path = Path(__file__).parent / 'prompts'
-prompt_name_question = 'Llama-2_QA_1.json'
-prompt_template_question = load_prompt(prompts_path / prompt_name_question)
-prompt_name_refine = 'Llama-2_QA-refine_1.json'
-prompt_template_refine = load_prompt(prompts_path / prompt_name_refine)
-
-
-def call_rag(query):
- retrieved_docs = retriever.get_chunk(query)
- sources = [doc.metadata["source"] for doc in retrieved_docs]
- responses = []
- for index, doc in enumerate(retrieved_docs):
- if index == 0:
- prompt = prompt_template_question.format(context=doc.page_content,
- question=query)
- response = llm.invoke(prompt)
- responses.append(response)
- else:
- prompt = prompt_template_refine.format(context=doc.page_content,
- question=query,
- existing_answer=responses[-1])
- response = llm.invoke(prompt)
- responses.append(response)
- return responses, sources
-
-
-if __name__ == "__main__":
- while True:
- query = input("Query:")
- responses, sources = call_rag(query)
- if not llm_conf['std_out']:
- print(responses[-1])
- print(f'Sources: ')
- for index, source in enumerate(sources):
- print(f'\t{index}: {source}')
diff --git a/projects/Retriver-GUI/retriever_app.py b/projects/Retriver-GUI/retriever_app.py
index f55c0c6..9f4198c 100644
--- a/projects/Retriver-GUI/retriever_app.py
+++ b/projects/Retriver-GUI/retriever_app.py
@@ -46,7 +46,7 @@ def render_search_results(self):
st.write(result.metadata)
def check_connection(self):
- response = self.app.retriever.client.test_connection()
+ response = self.app.retriever.vectordb.test_connection()
if response:
return True
else:
@@ -55,14 +55,14 @@ def check_connection(self):
def render_stats(self):
st.write(f'''
**Chroma Client Details:** \n
- Host Address : {self.app.retriever.client.host}:{self.app.retriever.client.port} \n
- Collection Name : {self.app.retriever.client.collection_name} \n
- Embeddings Type : {self.app.retriever.client.embedding_type} \n
- Embeddings Model: {self.app.retriever.client.embedding_model} \n
- Number of docs : {self.app.retriever.client.collection.count()} \n
+ Host Address : {self.app.retriever.vectordb.host}:{self.app.retriever.vectordb.port} \n
+ Collection Name : {self.app.retriever.vectordb.collection_name} \n
+ Embeddings Type : {self.app.retriever.vectordb.embedding_type} \n
+ Embeddings Model: {self.app.retriever.vectordb.embedding_model} \n
+ Number of docs : {self.app.retriever.vectordb.collection.count()} \n
''')
if st.button('Check Connection'):
- response = self.app.retriever.client.test_connection()
+ response = self.app.retriever.vectordb.test_connection()
if response:
st.write(':green[Connection Active]')
else:
diff --git a/pyproject.toml b/pyproject.toml
index f7c2d4f..b97ba19 100644
--- a/pyproject.toml
+++ b/pyproject.toml
@@ -42,6 +42,7 @@ dependencies = [
"huggingface_hub>=0.20.2",
"pydantic>=2.5.0",
"rouge-score>=0.1.2",
+ "deeplake>=3.8.27"
]
[project.urls]
diff --git a/src/config.ini b/src/config.ini
index 54990bf..c2938f9 100644
--- a/src/config.ini
+++ b/src/config.ini
@@ -14,6 +14,12 @@ n_gpu_layers_cpp : -1
std_out : True
base_dir : ${root:root_path}/models
+[deeplake]
+collection_name : arxiv
+embedding_type : instructor-embedding
+embedding_model : hkunlp/instructor-xl
+store_path : ${data:data_path}/vectordb
+
[chroma]
host : localhost
port : 8000
@@ -59,4 +65,7 @@ table_as_html : True
data_path : ${root:root_path}/data
[root]
-root_path : /home/ubuntu/CapStone/Capstone_5
+root_path : /home/ubuntu/volume_2k/Capstone_5
+
+[quantize]
+llama_cpp_path : ${root:root_path}
diff --git a/src/grag/components/chroma_client.py b/src/grag/components/chroma_client.py
index 7efd7c3..a03caa9 100644
--- a/src/grag/components/chroma_client.py
+++ b/src/grag/components/chroma_client.py
@@ -1,3 +1,4 @@
+import asyncio
from typing import List
import chromadb
@@ -7,13 +8,14 @@
from langchain_community.vectorstores.utils import filter_complex_metadata
from langchain_core.documents import Document
from tqdm import tqdm
-from tqdm.asyncio import tqdm as atqdm
+from tqdm.asyncio import tqdm_asyncio
chroma_conf = get_config()["chroma"]
class ChromaClient:
- """A class for connecting to a hosted Chroma Vectorstore collection.
+ """
+ A class for connecting to a hosted Chroma Vectorstore collection.
Attributes:
host : str
@@ -44,13 +46,15 @@ def __init__(
embedding_type=chroma_conf["embedding_type"],
embedding_model=chroma_conf["embedding_model"],
):
- """Args:
- host: IP Address of hosted Chroma Vectorstore, defaults to argument from config file
- port: port address of hosted Chroma Vectorstore, defaults to argument from config file
- collection_name: name of the collection in the Chroma Vectorstore, defaults to argument from config file
- embedding_type: type of embedding used, supported 'sentence-transformers' and 'instructor-embedding', defaults to argument from config file
- embedding_model: model name of embedding used, should correspond to the embedding_type, defaults to argument from config file
"""
+ Args:
+ host: IP Address of hosted Chroma Vectorstore, defaults to argument from config file
+ port: port address of hosted Chroma Vectorstore, defaults to argument from config file
+ collection_name: name of the collection in the Chroma Vectorstore, defaults to argument from config file
+ embedding_type: type of embedding used, supported 'sentence-transformers' and 'instructor-embedding', defaults to argument from config file
+ embedding_model: model name of embedding used, should correspond to the embedding_type, defaults to argument from config file
+ """
+
self.host: str = host
self.port: str = port
self.collection_name: str = collection_name
@@ -73,7 +77,8 @@ def __init__(
self.allowed_metadata_types = (str, int, float, bool)
def test_connection(self, verbose=True):
- """Tests connection with Chroma Vectorstore
+ """
+ Tests connection with Chroma Vectorstore
Args:
verbose: if True, prints connection status
@@ -90,7 +95,8 @@ def test_connection(self, verbose=True):
return response
async def aadd_docs(self, docs: List[Document], verbose=True):
- """Asynchronously adds documents to chroma vectorstore
+ """
+ Asynchronously adds documents to chroma vectorstore
Args:
docs: List of Documents
@@ -100,26 +106,17 @@ async def aadd_docs(self, docs: List[Document], verbose=True):
None
"""
docs = self._filter_metadata(docs)
- # tasks = [self.langchain_chroma.aadd_documents([doc]) for doc in docs]
- # if verbose:
- # await tqdm_asyncio.gather(*tasks, desc=f'Adding to {self.collection_name}')
- # else:
- # await asyncio.gather(*tasks)
+ tasks = [self.langchain_chroma.aadd_documents([doc]) for doc in docs]
if verbose:
- for doc in atqdm(
- docs,
- desc=f"Adding documents to {self.collection_name}",
- total=len(docs),
- ):
- await self.langchain_chroma.aadd_documents([doc])
+ await tqdm_asyncio.gather(*tasks, desc=f"Adding to {self.collection_name}")
else:
- for doc in docs:
- await self.langchain_chroma.aadd_documents([doc])
+ await asyncio.gather(*tasks)
def add_docs(self, docs: List[Document], verbose=True):
- """Adds documents to chroma vectorstore
+ """
+ Adds documents to chroma vectorstore
- Args:
+ Args:
docs: List of Documents
verbose: Show progress bar
diff --git a/src/grag/components/multivec_retriever.py b/src/grag/components/multivec_retriever.py
index 97684dd..05478df 100644
--- a/src/grag/components/multivec_retriever.py
+++ b/src/grag/components/multivec_retriever.py
@@ -6,14 +6,19 @@
import asyncio
import uuid
-from typing import List
+from pathlib import Path
+from typing import Any, Dict, List, Optional, Union
-from grag.components.chroma_client import ChromaClient
+from grag.components.parse_pdf import ParsePDF
from grag.components.text_splitter import TextSplitter
from grag.components.utils import get_config
+from grag.components.vectordb.base import VectorDB
+from grag.components.vectordb.deeplake_client import DeepLakeClient
from langchain.retrievers.multi_vector import MultiVectorRetriever
from langchain.storage import LocalFileStore
from langchain_core.documents import Document
+from tqdm import tqdm
+from tqdm.asyncio import tqdm as atqdm
multivec_retriever_conf = get_config()["multivec_retriever"]
@@ -28,9 +33,10 @@ class Retriever:
Attributes:
store_path: Path to the local file store
id_key: A key prefix for identifying documents
- client: ChromaClient class instance from components.chroma_client
+ vectordb: ChromaClient class instance from components.client
store: langchain.storage.LocalFileStore object, stores the key value pairs of document id and parent file
- retriever: langchain.retrievers.multi_vector.MultiVectorRetriever class instance, langchain's multi-vector retriever
+ retriever: langchain.retrievers.multi_vector.MultiVectorRetriever class instance,
+ langchain's multi-vector retriever
splitter: TextSplitter class instance from components.text_splitter
namespace: Namespace for producing unique id
top_k: Number of top chunks to return from similarity search.
@@ -39,26 +45,36 @@ class Retriever:
def __init__(
self,
+ vectordb: Optional[VectorDB] = None,
store_path: str = multivec_retriever_conf["store_path"],
id_key: str = multivec_retriever_conf["id_key"],
namespace: str = multivec_retriever_conf["namespace"],
top_k=1,
+ client_kwargs: Optional[Dict[str, Any]] = None,
):
"""Initialize the Retriever.
Args:
+ vectordb: Vector DB client instance
store_path: Path to the local file store, defaults to argument from config file
id_key: A key prefix for identifying documents, defaults to argument from config file
namespace: A namespace for producing unique id, defaults to argument from congig file
top_k: Number of top chunks to return from similarity search, defaults to 1
+ client_kwargs: kwargs to pass to the vectordb client
"""
self.store_path = store_path
self.id_key = id_key
self.namespace = uuid.UUID(namespace)
- self.client = ChromaClient()
+ if vectordb is None:
+ if client_kwargs is not None:
+ self.vectordb = DeepLakeClient(**client_kwargs)
+ else:
+ self.vectordb = DeepLakeClient()
+ else:
+ self.vectordb = vectordb
self.store = LocalFileStore(self.store_path)
self.retriever = MultiVectorRetriever(
- vectorstore=self.client.langchain_chroma,
+ vectorstore=self.vectordb.langchain_client,
byte_store=self.store,
id_key=self.id_key,
)
@@ -125,7 +141,7 @@ def add_docs(self, docs: List[Document]):
"""
chunks = self.split_docs(docs)
doc_ids = self.gen_doc_ids(docs)
- self.client.add_docs(chunks)
+ self.vectordb.add_docs(chunks)
self.retriever.docstore.mset(list(zip(doc_ids, docs)))
async def aadd_docs(self, docs: List[Document]):
@@ -140,11 +156,11 @@ async def aadd_docs(self, docs: List[Document]):
"""
chunks = self.split_docs(docs)
doc_ids = self.gen_doc_ids(docs)
- await asyncio.run(self.client.aadd_docs(chunks))
+ await asyncio.run(self.vectordb.aadd_docs(chunks))
self.retriever.docstore.mset(list(zip(doc_ids)))
def get_chunk(self, query: str, with_score=False, top_k=None):
- """Returns the most (cosine) similar chunks from the vector database.
+ """Returns the most similar chunks from the vector database.
Args:
query: A query string
@@ -155,14 +171,8 @@ def get_chunk(self, query: str, with_score=False, top_k=None):
list of Documents
"""
- if with_score:
- return self.client.langchain_chroma.similarity_search_with_relevance_scores(
- query=query, **{"k": top_k} if top_k else self.retriever.search_kwargs
- )
- else:
- return self.client.langchain_chroma.similarity_search(
- query=query, **{"k": top_k} if top_k else self.retriever.search_kwargs
- )
+ _top_k = top_k if top_k else self.retriever.search_kwargs["k"]
+ return self.vectordb.get_chunk(query=query, top_k=_top_k, with_score=with_score)
async def aget_chunk(self, query: str, with_score=False, top_k=None):
"""Returns the most (cosine) similar chunks from the vector database, asynchronously.
@@ -176,14 +186,10 @@ async def aget_chunk(self, query: str, with_score=False, top_k=None):
list of Documents
"""
- if with_score:
- return await self.client.langchain_chroma.asimilarity_search_with_relevance_scores(
- query=query, **{"k": top_k} if top_k else self.retriever.search_kwargs
- )
- else:
- return await self.client.langchain_chroma.asimilarity_search(
- query=query, **{"k": top_k} if top_k else self.retriever.search_kwargs
- )
+ _top_k = top_k if top_k else self.retriever.search_kwargs["k"]
+ return await self.vectordb.aget_chunk(
+ query=query, top_k=_top_k, with_score=with_score
+ )
def get_doc(self, query: str):
"""Returns the parent document of the most (cosine) similar chunk from the vector database.
@@ -228,3 +234,87 @@ def get_docs_from_chunks(self, chunks: List[Document], one_to_one=False):
ids.append(d.metadata[self.id_key])
docs = self.retriever.docstore.mget(ids)
return [d for d in docs if d is not None]
+
+ def ingest(
+ self,
+ dir_path: Union[str, Path],
+ glob_pattern: str = "**/*.pdf",
+ dry_run: bool = False,
+ verbose: bool = True,
+ parser_kwargs: dict = None,
+ ):
+ """Ingests the files in directory.
+
+ Args:
+ dir_path: path to the directory
+ glob_pattern: glob pattern to identify files
+ dry_run: if True, does not ingest any files
+ verbose: if True, shows progress
+ parser_kwargs: arguments to pass to the parser
+
+ """
+ _formats_to_add = ["Text", "Tables"]
+ filepath_gen = Path(dir_path).glob(glob_pattern)
+ if parser_kwargs:
+ parser = ParsePDF(parser_kwargs)
+ else:
+ parser = ParsePDF()
+ if verbose:
+ num_files = len(list(Path(dir_path).glob(glob_pattern)))
+ pbar = tqdm(filepath_gen, total=num_files, desc="Ingesting Files")
+ for filepath in pbar:
+ if not dry_run:
+ pbar.set_postfix_str(
+ f"Parsing file - {filepath.relative_to(dir_path)}"
+ )
+ docs = parser.load_file(filepath)
+ pbar.set_postfix_str(
+ f"Adding file - {filepath.relative_to(dir_path)}"
+ )
+ for format_key in _formats_to_add:
+ self.add_docs(docs[format_key])
+ print(f"Completed adding - {filepath.relative_to(dir_path)}")
+ else:
+ print(f"DRY RUN: found - {filepath.relative_to(dir_path)}")
+
+ async def aingest(
+ self,
+ dir_path: Union[str, Path],
+ glob_pattern: str = "**/*.pdf",
+ dry_run: bool = False,
+ verbose: bool = True,
+ parser_kwargs: dict = None,
+ ):
+ """Asynchronously ingests the files in directory.
+
+ Args:
+ dir_path: path to the directory
+ glob_pattern: glob pattern to identify files
+ dry_run: if True, does not ingest any files
+ verbose: if True, shows progress
+ parser_kwargs: arguments to pass to the parser
+
+ """
+ _formats_to_add = ["Text", "Tables"]
+ filepath_gen = Path(dir_path).glob(glob_pattern)
+ if parser_kwargs:
+ parser = ParsePDF(parser_kwargs)
+ else:
+ parser = ParsePDF()
+ if verbose:
+ num_files = len(list(Path(dir_path).glob(glob_pattern)))
+ pbar = atqdm(filepath_gen, total=num_files, desc="Ingesting Files")
+ for filepath in pbar:
+ if not dry_run:
+ pbar.set_postfix_str(
+ f"Parsing file - {filepath.relative_to(dir_path)}"
+ )
+ docs = parser.load_file(filepath)
+ pbar.set_postfix_str(
+ f"Adding file - {filepath.relative_to(dir_path)}"
+ )
+ for format_key in _formats_to_add:
+ await self.aadd_docs(docs[format_key])
+ print(f"Completed adding - {filepath.relative_to(dir_path)}")
+ else:
+ print(f"DRY RUN: found - {filepath.relative_to(dir_path)}")
diff --git a/src/grag/components/vectordb/__init__.py b/src/grag/components/vectordb/__init__.py
new file mode 100644
index 0000000..e69de29
diff --git a/src/grag/components/vectordb/base.py b/src/grag/components/vectordb/base.py
new file mode 100644
index 0000000..b0b0623
--- /dev/null
+++ b/src/grag/components/vectordb/base.py
@@ -0,0 +1,85 @@
+"""Abstract base class for vector database clients.
+
+This module provides:
+- VectorDB
+"""
+
+from abc import ABC, abstractmethod
+from typing import List, Tuple, Union
+
+from langchain_community.vectorstores.utils import filter_complex_metadata
+from langchain_core.documents import Document
+
+
+class VectorDB(ABC):
+ """Abstract base class for vector database clients."""
+
+ @abstractmethod
+ def __len__(self) -> int:
+ """Number of chunks in the vector database."""
+ ...
+
+ @abstractmethod
+ def delete(self) -> None:
+ """Delete all chunks in the vector database."""
+
+ @abstractmethod
+ def add_docs(self, docs: List[Document], verbose: bool = True) -> None:
+ """Adds documents to the vector database.
+
+ Args:
+ docs: List of Documents
+ verbose: Show progress bar
+
+ Returns:
+ None
+ """
+ ...
+
+ @abstractmethod
+ async def aadd_docs(self, docs: List[Document], verbose: bool = True) -> None:
+ """Adds documents to the vector database (asynchronous).
+
+ Args:
+ docs: List of Documents
+ verbose: Show progress bar
+
+ Returns:
+ None
+ """
+ ...
+
+ @abstractmethod
+ def get_chunk(
+ self, query: str, with_score: bool = False, top_k: int = None
+ ) -> Union[List[Document], List[Tuple[Document, float]]]:
+ """Returns the most similar chunks from the vector database.
+
+ Args:
+ query: A query string
+ with_score: Outputs scores of returned chunks
+ top_k: Number of top similar chunks to return, if None defaults to self.top_k
+
+ Returns:
+ list of Documents
+ """
+ ...
+
+ @abstractmethod
+ async def aget_chunk(
+ self, query: str, with_score: bool = False, top_k: int = None
+ ) -> Union[List[Document], List[Tuple[Document, float]]]:
+ """Returns the most similar chunks from the vector database (asynchronous).
+
+ Args:
+ query: A query string
+ with_score: Outputs scores of returned chunks
+ top_k: Number of top similar chunks to return, if None defaults to self.top_k
+
+ Returns:
+ list of Documents
+ """
+ ...
+
+ def _filter_metadata(self, docs: List[Document]) -> List[Document]:
+ return filter_complex_metadata(docs, allowed_types=self.allowed_metadata_types)
diff --git a/src/grag/components/vectordb/chroma_client.py b/src/grag/components/vectordb/chroma_client.py
new file mode 100644
index 0000000..cac8ab3
--- /dev/null
+++ b/src/grag/components/vectordb/chroma_client.py
@@ -0,0 +1,198 @@
+"""Class for Chroma vector database.
+
+This module provides:
+- ChromaClient
+"""
+
+from typing import List, Tuple, Union
+
+import chromadb
+from grag.components.embedding import Embedding
+from grag.components.utils import get_config
+from grag.components.vectordb.base import VectorDB
+from langchain_community.vectorstores import Chroma
+from langchain_core.documents import Document
+from tqdm import tqdm
+from tqdm.asyncio import tqdm as atqdm
+
+chroma_conf = get_config()["chroma"]
+
+
+class ChromaClient(VectorDB):
+ """A class for connecting to a hosted Chroma Vectorstore collection.
+
+ Attributes:
+ host : str
+ IP Address of hosted Chroma Vectorstore
+ port : str
+ port address of hosted Chroma Vectorstore
+ collection_name : str
+ name of the collection in the Chroma Vectorstore, each ChromaClient connects to a single collection
+ embedding_type : str
+ type of embedding used, supported 'sentence-transformers' and 'instructor-embedding'
+ embedding_model : str
+ model name of embedding used, should correspond to the embedding_type
+ embedding_function
+ a function of the embedding model, derived from the embedding_type and embedding_modelname
+ client: chromadb.HttpClient
+ Chroma API for client
+ collection
+ Chroma API for the collection
+ langchain_client: langchain_community.vectorstores.Chroma
+ LangChain wrapper for Chroma collection
+ """
+
+ def __init__(
+ self,
+ host=chroma_conf["host"],
+ port=chroma_conf["port"],
+ collection_name=chroma_conf["collection_name"],
+ embedding_type=chroma_conf["embedding_type"],
+ embedding_model=chroma_conf["embedding_model"],
+ ):
+ """Initialize a ChromaClient object.
+
+ Args:
+ host: IP Address of hosted Chroma Vectorstore, defaults to argument from config file
+ port: port address of hosted Chroma Vectorstore, defaults to argument from config file
+ collection_name: name of the collection in the Chroma Vectorstore, defaults to argument from config file
+ embedding_type: type of embedding used, supported 'sentence-transformers' and 'instructor-embedding', defaults to argument from config file
+ embedding_model: model name of embedding used, should correspond to the embedding_type, defaults to argument from config file
+ """
+ self.host: str = host
+ self.port: str = port
+ self.collection_name: str = collection_name
+ self.embedding_type: str = embedding_type
+ self.embedding_model: str = embedding_model
+
+ self.embedding_function = Embedding(
+ embedding_model=self.embedding_model, embedding_type=self.embedding_type
+ ).embedding_function
+
+ self.client = chromadb.HttpClient(host=self.host, port=self.port)
+ self.collection = self.client.get_or_create_collection(
+ name=self.collection_name
+ )
+ self.langchain_client = Chroma(
+ client=self.client,
+ collection_name=self.collection_name,
+ embedding_function=self.embedding_function,
+ )
+ self.allowed_metadata_types = (str, int, float, bool)
+
+ def __len__(self) -> int:
+ """Count the number of chunks in the database."""
+ return self.collection.count()
+
+ def delete(self) -> None:
+ """Delete all the chunks in the database collection."""
+ self.client.delete_collection(self.collection_name)
+ self.collection = self.client.get_or_create_collection(
+ name=self.collection_name
+ )
+ self.langchain_client = Chroma(
+ client=self.client,
+ collection_name=self.collection_name,
+ embedding_function=self.embedding_function,
+ )
+
+ def test_connection(self, verbose=True) -> int:
+ """Tests connection with Chroma Vectorstore.
+
+ Args:
+ verbose: if True, prints connection status
+
+ Returns:
+ A random integer if connection is alive else None
+ """
+ response = self.client.heartbeat()
+ if verbose:
+ if response:
+ print(f"Connection to {self.host}/{self.port} is alive..")
+ else:
+ print(f"Connection to {self.host}/{self.port} is not alive !!")
+ return response
+
+ def add_docs(self, docs: List[Document], verbose=True) -> None:
+ """Adds documents to chroma vectorstore.
+
+ Args:
+ docs: List of Documents
+ verbose: Show progress bar
+
+ Returns:
+ None
+ """
+ docs = self._filter_metadata(docs)
+ for doc in (
+ tqdm(docs, desc=f"Adding to {self.collection_name}:") if verbose else docs
+ ):
+ _id = self.langchain_client.add_documents([doc])
+
+ async def aadd_docs(self, docs: List[Document], verbose=True) -> None:
+ """Asynchronously adds documents to chroma vectorstore.
+
+ Args:
+ docs: List of Documents
+ verbose: Show progress bar
+
+ Returns:
+ None
+ """
+ docs = self._filter_metadata(docs)
+ if verbose:
+ for doc in atqdm(
+ docs,
+ desc=f"Adding documents to {self.collection_name}",
+ total=len(docs),
+ ):
+ await self.langchain_client.aadd_documents([doc])
+ else:
+ for doc in docs:
+ await self.langchain_client.aadd_documents([doc])
+
+ def get_chunk(
+ self, query: str, with_score: bool = False, top_k: int = None
+ ) -> Union[List[Document], List[Tuple[Document, float]]]:
+ """Returns the most similar chunks from the chroma database.
+
+ Args:
+ query: A query string
+ with_score: Outputs scores of returned chunks
+ top_k: Number of top similar chunks to return, if None defaults to self.top_k
+
+ Returns:
+ list of Documents
+
+ """
+ if with_score:
+ return self.langchain_client.similarity_search_with_relevance_scores(
+ query=query, k=top_k if top_k else 1
+ )
+ else:
+ return self.langchain_client.similarity_search(
+ query=query, k=top_k if top_k else 1
+ )
+
+ async def aget_chunk(
+ self, query: str, with_score=False, top_k=None
+ ) -> Union[List[Document], List[Tuple[Document, float]]]:
+ """Returns the most (cosine) similar chunks from the vector database, asynchronously.
+
+ Args:
+ query: A query string
+ with_score: Outputs scores of returned chunks
+ top_k: Number of top similar chunks to return, if None defaults to self.top_k
+
+ Returns:
+ list of Documents
+
+ """
+ if with_score:
+ return await self.langchain_client.asimilarity_search_with_relevance_scores(
+ query=query, k=top_k if top_k else 1
+ )
+ else:
+ return await self.langchain_client.asimilarity_search(
+ query=query, k=top_k if top_k else 1
+ )
diff --git a/src/grag/components/vectordb/deeplake_client.py b/src/grag/components/vectordb/deeplake_client.py
new file mode 100644
index 0000000..f0d5ba5
--- /dev/null
+++ b/src/grag/components/vectordb/deeplake_client.py
@@ -0,0 +1,159 @@
+"""Class for DeepLake vector database.
+
+This module provides:
+- DeepLakeClient
+"""
+
+from pathlib import Path
+from typing import List, Tuple, Union
+
+from grag.components.embedding import Embedding
+from grag.components.utils import get_config
+from grag.components.vectordb.base import VectorDB
+from langchain_community.vectorstores import DeepLake
+from langchain_core.documents import Document
+from tqdm import tqdm
+from tqdm.asyncio import tqdm as atqdm
+
+deeplake_conf = get_config()["deeplake"]
+
+
+class DeepLakeClient(VectorDB):
+ """A class for connecting to a DeepLake Vectorstore.
+
+ Attributes:
+ store_path : str, Path
+ The path to store the DeepLake vectorstore.
+ embedding_type : str
+ type of embedding used, supported 'sentence-transformers' and 'instructor-embedding'
+ embedding_model : str
+ model name of embedding used, should correspond to the embedding_type
+ embedding_function
+ a function of the embedding model, derived from the embedding_type and embedding_modelname
+ client: deeplake.core.vectorstore.VectorStore
+ DeepLake API
+ collection
+ Chroma API for the collection
+ langchain_client: langchain_community.vectorstores.DeepLake
+ LangChain wrapper for DeepLake API
+ """
+
+ def __init__(
+ self,
+ collection_name: str = deeplake_conf["collection_name"],
+ store_path: Union[str, Path] = deeplake_conf["store_path"],
+ embedding_type: str = deeplake_conf["embedding_type"],
+ embedding_model: str = deeplake_conf["embedding_model"],
+ read_only: bool = False,
+ ):
+ """Initialize DeepLake client object."""
+ self.store_path = Path(store_path)
+ self.collection_name = collection_name
+ self.read_only = read_only
+ self.embedding_type: str = embedding_type
+ self.embedding_model: str = embedding_model
+
+ self.embedding_function = Embedding(
+ embedding_model=self.embedding_model, embedding_type=self.embedding_type
+ ).embedding_function
+
+ # self.client = VectorStore(path=self.store_path / self.collection_name)
+ self.langchain_client = DeepLake(
+ dataset_path=str(self.store_path / self.collection_name),
+ embedding=self.embedding_function,
+ read_only=self.read_only,
+ )
+ self.client = self.langchain_client.vectorstore
+ self.allowed_metadata_types = (str, int, float, bool)
+
+ def __len__(self) -> int:
+ """Number of chunks in the vector database."""
+ return self.client.__len__()
+
+ def delete(self) -> None:
+ """Delete all chunks in the vector database."""
+ self.client.delete(delete_all=True)
+
+ def add_docs(self, docs: List[Document], verbose=True) -> None:
+ """Adds documents to deeplake vectorstore.
+
+ Args:
+ docs: List of Documents
+ verbose: Show progress bar
+
+ Returns:
+ None
+ """
+ docs = self._filter_metadata(docs)
+ for doc in (
+ tqdm(docs, desc=f"Adding to {self.collection_name}:") if verbose else docs
+ ):
+ _id = self.langchain_client.add_documents([doc])
+
+ async def aadd_docs(self, docs: List[Document], verbose=True) -> None:
+ """Asynchronously adds documents to chroma vectorstore.
+
+ Args:
+ docs: List of Documents
+ verbose: Show progress bar
+
+ Returns:
+ None
+ """
+ docs = self._filter_metadata(docs)
+ if verbose:
+ for doc in atqdm(
+ docs,
+ desc=f"Adding documents to {self.collection_name}",
+ total=len(docs),
+ ):
+ await self.langchain_client.aadd_documents([doc])
+ else:
+ for doc in docs:
+ await self.langchain_client.aadd_documents([doc])
+
+ def get_chunk(
+ self, query: str, with_score: bool = False, top_k: int = None
+ ) -> Union[List[Document], List[Tuple[Document, float]]]:
+ """Returns the most similar chunks from the deeplake database.
+
+ Args:
+ query: A query string
+ with_score: Outputs scores of returned chunks
+ top_k: Number of top similar chunks to return, if None defaults to self.top_k
+
+ Returns:
+ list of Documents
+
+ """
+ if with_score:
+ return self.langchain_client.similarity_search_with_score(
+ query=query, k=top_k if top_k else 1
+ )
+ else:
+ return self.langchain_client.similarity_search(
+ query=query, k=top_k if top_k else 1
+ )
+
+ async def aget_chunk(
+ self, query: str, with_score=False, top_k=None
+ ) -> Union[List[Document], List[Tuple[Document, float]]]:
+ """Returns the most similar chunks from the deeplake database, asynchronously.
+
+ Args:
+ query: A query string
+ with_score: Outputs scores of returned chunks
+ top_k: Number of top similar chunks to return, if None defaults to self.top_k
+
+ Returns:
+ list of Documents
+
+ """
+ if with_score:
+ return await self.langchain_client.asimilarity_search_with_score(
+ query=query, k=top_k if top_k else 1
+ )
+ else:
+ return await self.langchain_client.asimilarity_search(
+ query=query, k=top_k if top_k else 1
+ )
diff --git a/src/grag/quantize/__init__.py b/src/grag/quantize/__init__.py
new file mode 100644
index 0000000..e69de29
diff --git a/src/grag/quantize/quantize.py b/src/grag/quantize/quantize.py
new file mode 100644
index 0000000..64fba47
--- /dev/null
+++ b/src/grag/quantize/quantize.py
@@ -0,0 +1,52 @@
+"""Interactive file for quantizing models."""
+
+from pathlib import Path
+
+from grag.components.utils import get_config
+from grag.quantize.utils import (
+ building_llamacpp,
+ fetch_model_repo,
+ get_llamacpp_repo,
+ quantize_model,
+)
+
+config = get_config()
+root_path = Path(config["quantize"]["llama_cpp_path"])
+
+if __name__ == "__main__":
+ user_input = input(
+ "Enter the path to the llama_cpp cloned repo, or where you'd like to clone it. Press Enter to use the default config path: "
+ ).strip()
+
+ if user_input != "":
+ root_path = Path(user_input)
+
+ res = get_llamacpp_repo(root_path)
+
+ if "Already up to date." in str(res.stdout):
+ print("Repository is already up to date. Skipping build.")
+ else:
+ print("Updates found. Starting build...")
+ building_llamacpp(root_path)
+
+ response = (
+ input("Do you want us to download the model? (y/n) [Enter for yes]: ")
+ .strip()
+ .lower()
+ )
+ if response == "n":
+ print("Please copy the model folder to 'llama.cpp/models/' folder.")
+ _ = input("Enter if you have already copied the model:")
+ model_dir = Path(input("Enter the model directory name: "))
+ elif response == "y" or response == "":
+ repo_id = input(
+ "Please enter the repo_id for the model (you can check on https://huggingface.co/models): "
+ ).strip()
+ fetch_model_repo(repo_id, root_path)
+ # model_dir = repo_id.split('/')[1]
+ model_dir = root_path / "llama.cpp" / "models" / repo_id.split("/")[1]
+
+ quantization = input(
+ "Enter quantization, recommended - Q5_K_M or Q4_K_M for more check https://github.com/ggerganov/llama.cpp/blob/master/examples/quantize/quantize.cpp#L19 : "
+ )
+ quantize_model(model_dir, quantization, root_path)
diff --git a/src/grag/quantize/utils.py b/src/grag/quantize/utils.py
new file mode 100644
index 0000000..bc1d280
--- /dev/null
+++ b/src/grag/quantize/utils.py
@@ -0,0 +1,135 @@
+"""Utility functions for quantization."""
+
+import os
+import subprocess
+from pathlib import Path
+from typing import Optional, Union
+
+from grag.components.utils import get_config
+from huggingface_hub import snapshot_download
+
+config = get_config()
+
+
+def get_llamacpp_repo(root_path: Union[str, Path]) -> subprocess.CompletedProcess:
+ """Clones or pulls the llama.cpp repository into the specified root path.
+
+ Args:
+ root_path: The root directory where the llama.cpp repository will be cloned or updated.
+
+ Returns:
+ A subprocess.CompletedProcess instance containing the result of the git operation.
+ """
+ if os.path.exists(f"{root_path}/llama.cpp"):
+ print(f"Repo exists at: {root_path}/llama.cpp")
+ res = subprocess.run(
+ ["git", "-C", f"{root_path}/llama.cpp", "pull"],
+ check=True,
+ capture_output=True,
+ )
+ else:
+ res = subprocess.run(
+ [
+ "git",
+ "clone",
+ "https://github.com/ggerganov/llama.cpp.git",
+ f"{root_path}/llama.cpp",
+ ],
+ check=True,
+ capture_output=True,
+ )
+
+ return res
+
+
+def building_llamacpp(root_path: Union[str, Path]) -> None:
+ """Attempts to build the llama.cpp project using make or cmake.
+
+ Args:
+ root_path (str): The root directory where the llama.cpp project is located.
+ """
+ os.chdir(f"{root_path}/llama.cpp/")
+ try:
+ subprocess.run(["which", "make"], check=True, stdout=subprocess.DEVNULL)
+ subprocess.run(["make", "LLAMA_CUBLAS=1"], check=True)
+ print("Llama.cpp build successful.")
+ except subprocess.CalledProcessError:
+ try:
+ subprocess.run(["which", "cmake"], check=True, stdout=subprocess.DEVNULL)
+ subprocess.run(["mkdir", "build"], check=True)
+ subprocess.run(
+ [
+ "cd",
+ "build",
+ "&&",
+ "cmake",
+ "..",
+ "-DLLAMA_CUBLAS=ON",
+ "&&",
+ "cmake",
+ "--build",
+ ".",
+ "--config",
+ "Release",
+ ],
+ shell=True,
+ check=True,
+ )
+ print("Llama.cpp build successful.")
+ except subprocess.CalledProcessError:
+ print("Unable to build, cannot find make or cmake.")
+ finally:
+ os.chdir(
+ Path(__file__).parent
+ ) # Assuming you want to return to the root path after operation
+
+
+def fetch_model_repo(repo_id: str, root_path: Union[str, Path]) -> None:
+ """Download model from huggingface.co/models.
+
+ Args:
+ repo_id (str): Repository ID of the model to download.
+ root_path (str): The root path where the model should be downloaded or copied.
+ """
+ local_dir = f"{root_path}/llama.cpp/models/{repo_id.split('/')[1]}"
+ os.makedirs(local_dir, exist_ok=True)
+ snapshot_download(
+ repo_id=repo_id,
+ local_dir=local_dir,
+ local_dir_use_symlinks="auto",
+ resume_download=True,
+ )
+ print(f"Model downloaded in {local_dir}")
+
+
+def quantize_model(
+ model_dir_path: Union[str, Path],
+ quantization: str,
+ root_path: Union[str, Path],
+ output_dir: Optional[Union[str, Path]] = None,
+) -> None:
+ """Quantizes a specified model using a given quantization level.
+
+ Args:
+ output_dir (str, Path, optional): Directory to save quantized model. Defaults to None
+ model_dir_path (str, Path): The directory path of the model to be quantized.
+ quantization (str): The quantization level to apply.
+ root_path (str, Path): The root directory path of the project.
+ """
+ os.chdir(f"{root_path}/llama.cpp/")
+ model_dir_path = Path(model_dir_path)
+ if output_dir is None:
+ output_dir = config["llm"]["base_dir"]
+
+ output_dir = Path(output_dir) / model_dir_path.name
+ os.makedirs(output_dir, exist_ok=True)
+
+ subprocess.run(["python3", "convert.py", f"{model_dir_path}/"], check=True)
+ model_file = model_dir_path / "ggml-model-f32.gguf"
+ quantized_model_file = output_dir / f"ggml-model-{quantization}.gguf"
+ subprocess.run(
+ ["./quantize", str(model_file), str(quantized_model_file), quantization],
+ check=True,
+ )
+ print(f"Quantized model present at {output_dir}")
+ os.chdir(Path(__file__).parent) # Return to the root path after operation
diff --git a/src/grag/rag/basic_rag.py b/src/grag/rag/basic_rag.py
index be055ba..da461b6 100644
--- a/src/grag/rag/basic_rag.py
+++ b/src/grag/rag/basic_rag.py
@@ -5,7 +5,7 @@
"""
import json
-from typing import List, Union
+from typing import List, Optional, Union
from grag import prompts
from grag.components.llm import LLM
@@ -32,20 +32,21 @@ class BasicRAG:
def __init__(
self,
+ retriever: Optional[Retriever] = None,
model_name=None,
doc_chain="stuff",
task="QA",
llm_kwargs=None,
retriever_kwargs=None,
- custom_prompt: Union[
- Prompt, FewShotPrompt, List[Prompt, FewShotPrompt], None
- ] = None,
+ custom_prompt: Union[Prompt, FewShotPrompt, None] = None,
):
- """Initialize BasicRAG."""
- if retriever_kwargs is None:
- self.retriever = Retriever()
+ if retriever is None:
+ if retriever_kwargs is None:
+ self.retriever = Retriever()
+ else:
+ self.retriever = Retriever(**retriever_kwargs)
else:
- self.retriever = Retriever(**retriever_kwargs)
+ self.retriever = retriever
if llm_kwargs is None:
self.llm_ = LLM()
diff --git a/src/tests/components/multivec_retriever_test.py b/src/tests/components/multivec_retriever_test.py
index 3ccb3fb..14dad0b 100644
--- a/src/tests/components/multivec_retriever_test.py
+++ b/src/tests/components/multivec_retriever_test.py
@@ -1,3 +1,92 @@
+import json
+
+from grag.components.multivec_retriever import Retriever
+from langchain_core.documents import Document
+
+retriever = Retriever() # pass test collection
+
+doc = Document(page_content="Hello worlds", metadata={"source": "bars"})
+
+
+def test_retriver_id_gen():
+ doc = Document(page_content="Hello world", metadata={"source": "bar"})
+ id_ = retriever.id_gen(doc)
+ assert isinstance(id, str)
+ assert len(id_) == 32
+ doc.page_content = doc.page_content + 'ABC'
+ id_1 = retriever.id_gen(doc)
+ assert id_ == id_1
+ doc.metadata["source"] = "bars"
+ id_1 = retriever.id_gen(doc)
+ assert id_ != id_1
+
+
+def test_retriever_gen_doc_ids():
+ docs = [Document(page_content="Hello world", metadata={"source": "bar"}),
+ Document(page_content="Hello", metadata={"source": "foo"})]
+ ids = retriever.gen_doc_ids(docs)
+ assert len(ids) == len(docs)
+ assert all(isinstance(id, str) for id in ids)
+
+
+def test_retriever_split_docs():
+ pass
+
+
+def test_retriever_split_docs():
+ pass
+
+
+def test_retriever_add_docs():
+ # small enough docs to not split.
+ docs = [Document(page_content=
+ """And so on this rainbow day, with storms all around them, and blue sky
+ above, they rode only as far as the valley. But from there, before they
+ turned to go back, the monuments appeared close, and they loomed
+ grandly with the background of purple bank and creamy cloud and shafts
+ of golden lightning. They seemed like sentinels--guardians of a great
+ and beautiful love born under their lofty heights, in the lonely
+ silence of day, in the star-thrown shadow of night. They were like that
+ love. And they held Lucy and Slone, calling every day, giving a
+ nameless and tranquil content, binding them true to love, true to the
+ sage and the open, true to that wild upland home""", metadata={"source": "test_doc_1"}),
+ Document(page_content=
+ """Slone and Lucy never rode down so far as the stately monuments, though
+ these held memories as hauntingly sweet as others were poignantly
+ bitter. Lucy never rode the King again. But Slone rode him, learned to
+ love him. And Lucy did not race any more. When Slone tried to stir in
+ her the old spirit all the response he got was a wistful shake of head
+ or a laugh that hid the truth or an excuse that the strain on her
+ ankles from Joel Creech's lasso had never mended. The girl was
+ unutterably happy, but it was possible that she would never race a
+ horse again.""", metadata={"source": "test_doc_2"}),
+ Document(page_content=
+ """Bostil wanted to be alone, to welcome the King, to lead him back to the
+ home corral, perhaps to hide from all eyes the change and the uplift
+ that would forever keep him from wronging another man.
+
+ The late rains came and like magic, in a few days, the sage grew green
+ and lustrous and fresh, the gray turning to purple.
+
+ Every morning the sun rose white and hot in a blue and cloudless sky.
+ And then soon the horizon line showed creamy clouds that rose and
+ spread and darkened. Every afternoon storms hung along the ramparts and
+ rainbows curved down beautiful and ethereal. The dim blackness of the
+ storm-clouds was split to the blinding zigzag of lightning, and the
+ thunder rolled and boomed, like the Colorado in flood.""", metadata={"source": "test_doc_3"})
+ ]
+ ids = retriever.gen_doc_ids(docs)
+ retriever.add_docs(docs)
+ retrieved = retriever.store.mget(ids)
+ assert len(retrieved) == len(ids)
+ for i, doc in enumerate(docs):
+ retrieved_doc = json.loads(retrieved[i].decode())
+ assert doc.metadata == retrieved_doc.metadata
+
+
+def test_retriever_aadd_docs():
+ pass
+
# # add code folder to sys path
# import os
# from pathlib import Path
diff --git a/src/tests/components/vectordb/__init__.py b/src/tests/components/vectordb/__init__.py
new file mode 100644
index 0000000..e69de29
diff --git a/src/tests/components/chroma_client_test.py b/src/tests/components/vectordb/chroma_client_test.py
similarity index 59%
rename from src/tests/components/chroma_client_test.py
rename to src/tests/components/vectordb/chroma_client_test.py
index 1596dd3..c491dfd 100644
--- a/src/tests/components/chroma_client_test.py
+++ b/src/tests/components/vectordb/chroma_client_test.py
@@ -1,12 +1,13 @@
import asyncio
-from grag.components.chroma_client import ChromaClient
+import pytest
+from grag.components.vectordb.chroma_client import ChromaClient
from langchain_core.documents import Document
def test_chroma_connection():
- client = ChromaClient()
- response = client.test_connection()
+ chroma_client = ChromaClient()
+ response = chroma_client.test_connection()
assert isinstance(response, int)
@@ -45,14 +46,12 @@ def test_chroma_add_docs():
storm-clouds was split to the blinding zigzag of lightning, and the
thunder rolled and boomed, like the Colorado in flood.""",
]
- client = ChromaClient(collection_name="test")
- if client.collection.count() > 0:
- client.chroma_client.delete_collection("test")
- client = ChromaClient(collection_name="test")
+ chroma_client = ChromaClient(collection_name="test")
+ if len(chroma_client) > 0:
+ chroma_client.delete()
docs = [Document(page_content=doc) for doc in docs]
- client.add_docs(docs)
- collection_count = client.collection.count()
- assert collection_count == len(docs)
+ chroma_client.add_docs(docs)
+ assert len(chroma_client) == len(docs)
def test_chroma_aadd_docs():
@@ -90,11 +89,60 @@ def test_chroma_aadd_docs():
storm-clouds was split to the blinding zigzag of lightning, and the
thunder rolled and boomed, like the Colorado in flood.""",
]
- client = ChromaClient(collection_name="test")
- if client.collection.count() > 0:
- client.chroma_client.delete_collection("test")
- client = ChromaClient(collection_name="test")
+ chroma_client = ChromaClient(collection_name="test")
+ if len(chroma_client) > 0:
+ chroma_client.delete()
docs = [Document(page_content=doc) for doc in docs]
loop = asyncio.get_event_loop()
- loop.run_until_complete(client.aadd_docs(docs))
- assert client.collection.count() == len(docs)
+ loop.run_until_complete(chroma_client.aadd_docs(docs))
+ assert len(chroma_client) == len(docs)
+
+
+chrome_get_chunk_params = [(1, False), (1, True), (2, False), (2, True)]
+
+
+@pytest.mark.parametrize("top_k,with_score", chrome_get_chunk_params)
+def test_chroma_get_chunk(top_k, with_score):
+ query = """Slone and Lucy never rode down so far as the stately monuments, though
+ these held memories as hauntingly sweet as others were poignantly
+ bitter. Lucy never rode the King again. But Slone rode him, learned to
+ love him. And Lucy did not race any more. When Slone tried to stir in
+ her the old spirit all the response he got was a wistful shake of head
+ or a laugh that hid the truth or an excuse that the strain on her
+ ankles from Joel Creech's lasso had never mended. The girl was
+ unutterably happy, but it was possible that she would never race a
+ horse again."""
+ chroma_client = ChromaClient(collection_name="test")
+ retrieved_chunks = chroma_client.get_chunk(
+ query=query, top_k=top_k, with_score=with_score
+ )
+ assert len(retrieved_chunks) == top_k
+ if with_score:
+ assert all(isinstance(doc[0], Document) for doc in retrieved_chunks)
+ assert all(isinstance(doc[1], float) for doc in retrieved_chunks)
+ else:
+ assert all(isinstance(doc, Document) for doc in retrieved_chunks)
+
+
+@pytest.mark.parametrize("top_k,with_score", chrome_get_chunk_params)
+def test_chroma_aget_chunk(top_k, with_score):
+ query = """Slone and Lucy never rode down so far as the stately monuments, though
+ these held memories as hauntingly sweet as others were poignantly
+ bitter. Lucy never rode the King again. But Slone rode him, learned to
+ love him. And Lucy did not race any more. When Slone tried to stir in
+ her the old spirit all the response he got was a wistful shake of head
+ or a laugh that hid the truth or an excuse that the strain on her
+ ankles from Joel Creech's lasso had never mended. The girl was
+ unutterably happy, but it was possible that she would never race a
+ horse again."""
+ chroma_client = ChromaClient(collection_name="test")
+ loop = asyncio.get_event_loop()
+ retrieved_chunks = loop.run_until_complete(
+ chroma_client.aget_chunk(query=query, top_k=top_k, with_score=with_score)
+ )
+ assert len(retrieved_chunks) == top_k
+ if with_score:
+ assert all(isinstance(doc[0], Document) for doc in retrieved_chunks)
+ assert all(isinstance(doc[1], float) for doc in retrieved_chunks)
+ else:
+ assert all(isinstance(doc, Document) for doc in retrieved_chunks)
diff --git a/src/tests/components/vectordb/deeplake_client_test.py b/src/tests/components/vectordb/deeplake_client_test.py
new file mode 100644
index 0000000..cea5e61
--- /dev/null
+++ b/src/tests/components/vectordb/deeplake_client_test.py
@@ -0,0 +1,146 @@
+import asyncio
+
+import pytest
+from grag.components.vectordb.deeplake_client import DeepLakeClient
+from langchain_core.documents import Document
+
+
+def test_deeplake_add_docs():
+ docs = [
+ """And so on this rainbow day, with storms all around them, and blue sky
+ above, they rode only as far as the valley. But from there, before they
+ turned to go back, the monuments appeared close, and they loomed
+ grandly with the background of purple bank and creamy cloud and shafts
+ of golden lightning. They seemed like sentinels--guardians of a great
+ and beautiful love born under their lofty heights, in the lonely
+ silence of day, in the star-thrown shadow of night. They were like that
+ love. And they held Lucy and Slone, calling every day, giving a
+ nameless and tranquil content, binding them true to love, true to the
+ sage and the open, true to that wild upland home.""",
+ """Slone and Lucy never rode down so far as the stately monuments, though
+ these held memories as hauntingly sweet as others were poignantly
+ bitter. Lucy never rode the King again. But Slone rode him, learned to
+ love him. And Lucy did not race any more. When Slone tried to stir in
+ her the old spirit all the response he got was a wistful shake of head
+ or a laugh that hid the truth or an excuse that the strain on her
+ ankles from Joel Creech's lasso had never mended. The girl was
+ unutterably happy, but it was possible that she would never race a
+ horse again.""",
+ """Bostil wanted to be alone, to welcome the King, to lead him back to the
+ home corral, perhaps to hide from all eyes the change and the uplift
+ that would forever keep him from wronging another man.
+
+ The late rains came and like magic, in a few days, the sage grew green
+ and lustrous and fresh, the gray turning to purple.
+
+ Every morning the sun rose white and hot in a blue and cloudless sky.
+ And then soon the horizon line showed creamy clouds that rose and
+ spread and darkened. Every afternoon storms hung along the ramparts and
+ rainbows curved down beautiful and ethereal. The dim blackness of the
+ storm-clouds was split to the blinding zigzag of lightning, and the
+ thunder rolled and boomed, like the Colorado in flood.""",
+ ]
+ deeplake_client = DeepLakeClient(collection_name="test")
+ if len(deeplake_client) > 0:
+ deeplake_client.delete()
+ docs = [Document(page_content=doc) for doc in docs]
+ deeplake_client.add_docs(docs)
+ assert len(deeplake_client) == len(docs)
+ del deeplake_client
+
+
+def test_chroma_aadd_docs():
+ docs = [
+ """And so on this rainbow day, with storms all around them, and blue sky
+ above, they rode only as far as the valley. But from there, before they
+ turned to go back, the monuments appeared close, and they loomed
+ grandly with the background of purple bank and creamy cloud and shafts
+ of golden lightning. They seemed like sentinels--guardians of a great
+ and beautiful love born under their lofty heights, in the lonely
+ silence of day, in the star-thrown shadow of night. They were like that
+ love. And they held Lucy and Slone, calling every day, giving a
+ nameless and tranquil content, binding them true to love, true to the
+ sage and the open, true to that wild upland home.""",
+ """Slone and Lucy never rode down so far as the stately monuments, though
+ these held memories as hauntingly sweet as others were poignantly
+ bitter. Lucy never rode the King again. But Slone rode him, learned to
+ love him. And Lucy did not race any more. When Slone tried to stir in
+ her the old spirit all the response he got was a wistful shake of head
+ or a laugh that hid the truth or an excuse that the strain on her
+ ankles from Joel Creech's lasso had never mended. The girl was
+ unutterably happy, but it was possible that she would never race a
+ horse again.""",
+ """Bostil wanted to be alone, to welcome the King, to lead him back to the
+ home corral, perhaps to hide from all eyes the change and the uplift
+ that would forever keep him from wronging another man.
+
+ The late rains came and like magic, in a few days, the sage grew green
+ and lustrous and fresh, the gray turning to purple.
+
+ Every morning the sun rose white and hot in a blue and cloudless sky.
+ And then soon the horizon line showed creamy clouds that rose and
+ spread and darkened. Every afternoon storms hung along the ramparts and
+ rainbows curved down beautiful and ethereal. The dim blackness of the
+ storm-clouds was split to the blinding zigzag of lightning, and the
+ thunder rolled and boomed, like the Colorado in flood.""",
+ ]
+ deeplake_client = DeepLakeClient(collection_name="test")
+ if len(deeplake_client) > 0:
+ deeplake_client.delete()
+ docs = [Document(page_content=doc) for doc in docs]
+ loop = asyncio.get_event_loop()
+ loop.run_until_complete(deeplake_client.aadd_docs(docs))
+ assert len(deeplake_client) == len(docs)
+ del deeplake_client
+
+
+deeplake_get_chunk_params = [(1, False), (1, True), (2, False), (2, True)]
+
+
+@pytest.mark.parametrize("top_k,with_score", deeplake_get_chunk_params)
+def test_deeplake_get_chunk(top_k, with_score):
+ query = """Slone and Lucy never rode down so far as the stately monuments, though
+ these held memories as hauntingly sweet as others were poignantly
+ bitter. Lucy never rode the King again. But Slone rode him, learned to
+ love him. And Lucy did not race any more. When Slone tried to stir in
+ her the old spirit all the response he got was a wistful shake of head
+ or a laugh that hid the truth or an excuse that the strain on her
+ ankles from Joel Creech's lasso had never mended. The girl was
+ unutterably happy, but it was possible that she would never race a
+ horse again."""
+ deeplake_client = DeepLakeClient(collection_name="test", read_only=True)
+ retrieved_chunks = deeplake_client.get_chunk(
+ query=query, top_k=top_k, with_score=with_score
+ )
+ assert len(retrieved_chunks) == top_k
+ if with_score:
+ assert all(isinstance(doc[0], Document) for doc in retrieved_chunks)
+ assert all(isinstance(doc[1], float) for doc in retrieved_chunks)
+ else:
+ assert all(isinstance(doc, Document) for doc in retrieved_chunks)
+ del deeplake_client
+
+
+@pytest.mark.parametrize("top_k,with_score", deeplake_get_chunk_params)
+def test_deeplake_aget_chunk(top_k, with_score):
+ query = """Slone and Lucy never rode down so far as the stately monuments, though
+ these held memories as hauntingly sweet as others were poignantly
+ bitter. Lucy never rode the King again. But Slone rode him, learned to
+ love him. And Lucy did not race any more. When Slone tried to stir in
+ her the old spirit all the response he got was a wistful shake of head
+ or a laugh that hid the truth or an excuse that the strain on her
+ ankles from Joel Creech's lasso had never mended. The girl was
+ unutterably happy, but it was possible that she would never race a
+ horse again."""
+ deeplake_client = DeepLakeClient(collection_name="test", read_only=True)
+ loop = asyncio.get_event_loop()
+ retrieved_chunks = loop.run_until_complete(
+ deeplake_client.aget_chunk(query=query, top_k=top_k, with_score=with_score)
+ )
+ assert len(retrieved_chunks) == top_k
+ if with_score:
+ assert all(isinstance(doc[0], Document) for doc in retrieved_chunks)
+ assert all(isinstance(doc[1], float) for doc in retrieved_chunks)
+ else:
+ assert all(isinstance(doc, Document) for doc in retrieved_chunks)
+ del deeplake_client
diff --git a/src/tests/quantize/__init__.py b/src/tests/quantize/__init__.py
new file mode 100644
index 0000000..e69de29
diff --git a/src/tests/quantize/quantize_test.py b/src/tests/quantize/quantize_test.py
new file mode 100644
index 0000000..af0e9dd
--- /dev/null
+++ b/src/tests/quantize/quantize_test.py
@@ -0,0 +1,39 @@
+import os
+from pathlib import Path
+
+from grag.quantize.utils import (
+ building_llamacpp,
+ fetch_model_repo,
+ get_llamacpp_repo,
+ quantize_model,
+)
+
+root_path = Path(__file__).parent / "test_data"
+os.makedirs(root_path, exist_ok=True)
+
+
+def test_get_llamacpp_repo():
+ get_llamacpp_repo(root_path)
+ repo_path = root_path / "llama.cpp" / ".git"
+ assert os.path.exists(repo_path)
+
+
+def test_build_llamacpp():
+ building_llamacpp(root_path)
+ bin_path = root_path / "llama.cpp" / "quantize"
+ assert os.path.exists(bin_path)
+
+
+def test_fetch_model_repo():
+ fetch_model_repo("meta-llama/Llama-2-7b-chat", root_path)
+ model_dir_path = root_path / "llama.cpp" / "models" / "Llama-2-7b-chat"
+ assert os.path.exists(model_dir_path)
+
+
+def test_quantize_model():
+ model_dir_path = root_path / "llama.cpp" / "models" / "Llama-2-7b-chat"
+ quantize_model(
+ model_dir_path, "Q3_K_M", root_path, output_dir=model_dir_path.parent
+ )
+ gguf_file_path = model_dir_path / "ggml-model-Q3_K_M.gguf"
+ assert os.path.exists(gguf_file_path)
diff --git a/src/tests/rag/basic_rag_test.py b/src/tests/rag/basic_rag_test.py
index 06db25e..2249028 100644
--- a/src/tests/rag/basic_rag_test.py
+++ b/src/tests/rag/basic_rag_test.py
@@ -1,4 +1,4 @@
-from typing import List, Text
+from typing import Text, List
from grag.rag.basic_rag import BasicRAG