Metadata-Version: 2.1
Name: mo-parsing
Version: 5.62.21283
Summary: Another PEG Parsing Tool
Home-page: https://github.com/klahnakoski/mo-parsing
Author: Various
Author-email: kyle@lahnakoski.com
License: MIT
Description: # More Parsing!
        
        A fork of [pyparsing](https://github.com/pyparsing/pyparsing) for faster parsing
        
        
        |Branch      |Status   |
        |------------|---------|
        |master      | [![Build Status](https://travis-ci.com/klahnakoski/mo-parsing.svg?branch=master)](https://travis-ci.com/klahnakoski/mo-parsing) |
        |dev         | [![Build Status](https://travis-ci.com/klahnakoski/mo-parsing.svg?branch=dev)](https://travis-ci.com/klahnakoski/mo-parsing)    |
        
        
        ## Differences from PyParsing
        
        This fork was originally created to support faster parsing for [mo-sql-parsing](https://github.com/klahnakoski/moz-sql-parser).  Since then it has deviated sufficiently to be it's own collection of parser specification functions.  Here are the differences:
        
        * Added `Whitespace`, which controls parsing context and whitespace.  It replaces the whitespace modifying methods of pyparsing
        * the wildcard ("`*`") could be used to indicate multi-values are expected; this is not allowed: all values are multi-values
        * all actions are in `f(token, index, string)` form, which is opposite of pyparsing's `f(string, index token)` form
        * ParserElements are static: For example, `expr.addParseAction(action)` creates a new ParserElement, so must be assigned to variable or it is lost. **This is the biggest source of bugs when converting from pyparsing**
        * removed all backward-compatibility settings
        * no support for binary serialization (no pickle)
        
        Faster Parsing
        
        * faster infix operator parsing (main reason for this fork)
        * ParseResults point to ParserElement for reduced size
        * regex used to reduce the number of failed parse attempts  
        * packrat parser is not need
        * less stack used 
        
        
        ## Details
        
        ### The `Whitespace` Skipper
        
        The `mo_parsing.whitespaces.CURRENT` is used during parser creation: It is effectively defines "whitespace" for skipping, with additional features to simplify the language definition.  You declare "standard" `Whitespace` like so:
        
            with Whitespace() as whitespace:
                # PUT YOUR LANGUAGE DEFINITION HERE (space, tab and CR are "whitespace")
        
        If you are declaring a large language, and you want to minimize indentation, and you are careful, you may also use this pattern:
        
            whitespace = Whitespace().use()
            # PUT YOUR LANGUAGE DEFINITION HERE
            whitespace.release()
        
        The whitespace can be used to set global parsing parameters, like
        
        * `set_whitespace()` - set the ignored characters (default: `"\t\n "`)
        * `add_ignore()` - include whole patterns that are ignored (like comments)
        * `set_literal()` - Set the definition for what `Literal()` means
        * `set_keyword_chars()` - For default `Keyword()` (important for defining word boundary)
        
        ### Navigating ParseResults
        
        The results off parsing are in `ParseResults` and are in the form of an n-ary tree; with the children found in `ParseResults.tokens`.  Each `ParseResult.type` points to the `ParserElement` that made it.  In general, if you want to get fancy with post processing (or in a `parseAction`), you will be required to navigate the raw `tokens` to generate a final result
        
        There are some convenience methods;  
        * `__iter__()` - allows you to iterate through parse results in **depth first search**. Empty results are skipped, and `Group`ed results are treated as atoms (which can be further iterated if required) 
        * `name` is a convenient property for `ParseResults.type.token_name`
        * `__getitem__()` - allows you to jump into the parse tree to the given `name`. This is blocked by any names found inside `Group`ed results (because groups are considered atoms).      
        
        ### addParseAction
        
        Parse actions are methods that are run after a ParserElement found a match. 
        
        * Parameters must be accepted in `(tokens, index, string)` order (the opposite of pyparsing)
        * Parse actions are wrapped to ensure the output is a legitimate ParseResult
          * If your parse action returns `None` then the result is the original `tokens`
          * If your parse action returns an object, or list, or tuple, then it will be packaged in a `ParseResult` with same type as `tokens`.
          * If your parse action returns a `ParseResult` then it is accepted ***even if is belongs to some other pattern***
          
        ### Debugging
        
        The PEG-style of mo-parsing (from pyparsing) makes a very expressible and readable specification, but debugging a parser is still hard.  To look deeper into what the parser is doing use the `Debugger`:
        
        ```
        with Debugger():
            expr.parseString("my new language")
        ```
        
        The debugger will print out details of what's happening
        
        * Each attempt, and if it matched or failed
        * A small number of bytes to show you the current position
        * location, line and character for more info about the current position
        * whitespace indicating stack depth
        * print out of the ParserElement performing the attempt
        
        This should help to to isolate the exact position your grammar is failing. 
        
        ## Contributing
        
        If you plan to extend or enhance this code, please [see the README in the tests directory](https://github.com/klahnakoski/mo-parsing/blob/dev/tests/README.md)
Platform: UNKNOWN
Classifier: Development Status :: 4 - Beta
Classifier: License :: OSI Approved :: MIT License
Classifier: Programming Language :: Python :: 3.7
Classifier: Topic :: Software Development :: Libraries
Classifier: Topic :: Software Development :: Libraries :: Python Modules
Description-Content-Type: text/markdown
Provides-Extra: tests
