diff --git a/.github/ISSUE_TEMPLATE/bug_report.md b/.github/ISSUE_TEMPLATE/bug_report.md new file mode 100644 index 000000000..087926a26 --- /dev/null +++ b/.github/ISSUE_TEMPLATE/bug_report.md @@ -0,0 +1,30 @@ +--- +name: Bug report +about: Create a report to help us improve + +--- + +**Describe the bug** +A clear and concise description of what the bug is. + +**To Reproduce** +Steps to reproduce the behavior: +1. import '...' +2. input '....' + +**Expected behavior** +A clear and concise description of what you expected to happen. + +**Source Code** +If applicable, add source code to help explain your problem. + +**Screenshots** +If applicable, add screenshots to help explain your problem. + +**Desktop (please complete the following information):** + - OS: [e.g. Windows,Linux] + - Python Version [e.g. 2.7, 3.6] + - Version [e.g. 22] + +**Additional context** +Add any other context about the problem here. diff --git a/.idea/codeStyles/codeStyleConfig.xml b/.idea/codeStyles/codeStyleConfig.xml new file mode 100644 index 000000000..8f1a3b79a --- /dev/null +++ b/.idea/codeStyles/codeStyleConfig.xml @@ -0,0 +1,5 @@ + + + + \ No newline at end of file diff --git a/.idea/markdown-navigator.xml b/.idea/markdown-navigator.xml new file mode 100644 index 000000000..4e63b3ef9 --- /dev/null +++ b/.idea/markdown-navigator.xml @@ -0,0 +1,78 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/.idea/markdown-navigator/profiles_settings.xml b/.idea/markdown-navigator/profiles_settings.xml new file mode 100644 index 000000000..57927c5a7 --- /dev/null +++ b/.idea/markdown-navigator/profiles_settings.xml @@ -0,0 +1,3 @@ + + + \ No newline at end of file diff --git a/.idea/misc.xml b/.idea/misc.xml new file mode 100644 index 000000000..3e9f8443e --- /dev/null +++ b/.idea/misc.xml @@ -0,0 +1,4 @@ + + + + \ No newline at end of file diff --git a/.idea/modules.xml b/.idea/modules.xml new file mode 100644 index 000000000..e1863b9d2 --- /dev/null +++ b/.idea/modules.xml @@ -0,0 +1,8 @@ + + + + + + + + \ No newline at end of file diff --git a/.idea/pythainlp.iml b/.idea/pythainlp.iml new file mode 100644 index 000000000..671160631 --- /dev/null +++ b/.idea/pythainlp.iml @@ -0,0 +1,11 @@ + + + + + + + + + + \ No newline at end of file diff --git a/.idea/vcs.xml b/.idea/vcs.xml new file mode 100644 index 000000000..94a25f7f4 --- /dev/null +++ b/.idea/vcs.xml @@ -0,0 +1,6 @@ + + + + + + \ No newline at end of file diff --git a/.idea/workspace.xml b/.idea/workspace.xml new file mode 100644 index 000000000..05c3ea8ee --- /dev/null +++ b/.idea/workspace.xml @@ -0,0 +1,586 @@ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ +
    + +
  • Docs »
  • + +
  • Overview: module code
  • + + +
  • + +
  • + +
+ + +
+
+ +
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ + + + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/change.html b/docs/_build/html/_modules/pythainlp/change.html new file mode 100644 index 000000000..44eb50101 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/change.html @@ -0,0 +1,268 @@ + + + + + + + + + + + pythainlp.change — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.change

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import,unicode_literals
+import six
+dictdata={u'Z':u'(',u'z':u'ผ',u'X':u')',u'x':u'ป',u'C':u'ฉ',u'c':u'แ',u'V':u'ฮ',u'v':u'อ',u'B':u'ฺ',u'b':u'ิ',u'N':u'์',u'n':u'ื',u'M':u'?',u'm':u'ท',u'<':u'ฒ',u',u':u'ม',u'>':u'ฬ',u'.':u'ใ',u'?':u'ฦ',u'/':u'ฝ',
+'A':u'ฤ',u'a':u'ฟ',u'S':u'ฆ',u's':u'ห',u'D':u'ฏ',u'd':u'ก',u'F':u'โ',u'f':u'ด',u'G':u'ฌ',u'g':u'เ',u'H':u'็',u'h':u'้',u'J':u'๋',u'j':u'j',u'K':u'ษ',u'k':u'า',u'L':u'ศ',u'l':u'ส',u':u':u'ซ',u'"':u'.',"'":"ง",u':u':u'ซ',u';':u'ว',
+'Q':u'๐',u'q':u'ๆ',u'W':u'"',u'w':u'ไ',u'E':u'ฎ',u'e':u'ำ',u'R':u'ฑ',u'r':u'พ',u'T':u'ธ',u't':u'ะ',u'Y':u'ํ',u'y':u'ั',u'U':u'๊',u'u':u'ี',u'I':u'ณ',u'i':u'ร',u'O':u'ฯ',u'o':u'น',u'P':u'ญ',u'p':u'ย',u'{':u'ฐ',u'[':u'บ',u'}':u',u',u']':u'ล',u'|':u'ฅ',u']':u'ฃ',
+'~':u'%',u'`':u'_',u'@':u'๑',u'2':u'/',u'#':u'๒',u'3':u'-',u'$':u'๓',u'4':u'ภ',u'%':u'๔',u'5':u'ถ',u'^':u'ู',u'6':u'ุ',u'&':u'฿',u'7':u'ึ',u'*':u'๕',u'8':u'ค',u'(':u'๖',u'9':u'ต',u')':u'๗',u'0':u'จ',u'_':u'๘',u'-':u'ข',u'+':u'๙',u'=':u'ช'}
+# แก้ไขพิมพ์ภาษาไทยผิดภาษา
+
[docs]def texttothai(data): + """ + :param str data: Incorrect input language correction (Needs thai but input english) + :return: thai text + """ + data = list(data) + data2 = "" + for a in data: + if a in dictdata: + a = dictdata[a] + else: + a = a + data2+=a + del data + return data2
+# แก้ไขพิมพ์ภาษาอังกฤษผิดภาษา +
[docs]def texttoeng(data): + """ + :param str data: Incorrect input language correction (Needs english but input thai) + :return: english text + """ + data = list(data) + data2 = "" + dictdataeng= {v: k for k, v in six.iteritems(dictdata)} + for a in data: + if a in dictdataeng: + a = dictdataeng[a] + else: + a = a + data2+=a + return data2
+if __name__ == "__main__": + a="l;ylfu8iy[" + a=texttothai(a) + a=texttothai(a) + b="นามรสนอำันี" + b=texttoeng(b) + six.print_(a) + six.print_(b) +
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/collation.html b/docs/_build/html/_modules/pythainlp/collation.html new file mode 100644 index 000000000..998707111 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/collation.html @@ -0,0 +1,251 @@ + + + + + + + + + + + pythainlp.collation — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.collation

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import, unicode_literals, print_function
+import re
+
+try:
+    import icu
+    thkey = icu.Collator.createInstance(icu.Locale('th_TH')).getSortKey
+except ImportError:
+    def thkey(word):
+        cv = re.sub('[็-์]', '', word,re.U) # remove tone
+        cv = re.sub('([เ-ไ])([ก-ฮ])', '\\2\\1', cv,re.U) # switch lead vowel
+        tone = re.sub('[^็-์]', ' ', word,re.U) # just tone
+        return cv+tone
+
+
[docs]def collation(data): + """ + :param list data: a list of thai text + :return: a list of thai text, sorted alphabetically + **Example**:: + >>> from pythainlp.collation import * + >>> collation(['ไก่', 'เป็ด', 'หมู', 'วัว']) + ['ไก่', 'เป็ด', 'วัว', 'หมู'] + """ + return sorted(data, key=thkey)
+ +if __name__ == "__main__": + a=collation(['ไก่','ไข่','ก','ฮา'])==['ก', 'ไก่', 'ไข่', 'ฮา'] + print(a) + print(collation(['หลาย','หญิง'])==['หญิง','หลาย']) + print(collation(['ไก่', 'เป็ด', 'หมู', 'วัว'])==['ไก่', 'เป็ด', 'วัว', 'หมู']) +
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/date.html b/docs/_build/html/_modules/pythainlp/date.html new file mode 100644 index 000000000..4b7eb1baf --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/date.html @@ -0,0 +1,262 @@ + + + + + + + + + + + pythainlp.date — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.date

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import,unicode_literals
+import datetime, pytz
+now1 = datetime.datetime.now()
+tz = pytz.timezone('Asia/Bangkok')
+
+
[docs]def now(): + """ + :return: the current date with Thai month and Thai year. The month is spelled out in text, and the year is converted from AD to Thai years. (ie: 30 ตุลาคม 2560 20:45:30) + """ + now1 = datetime.datetime.now(tz) + month_name = 'x มกราคม กุมภาพันธ์ มีนาคม เมษายน พฤษภาคม มิถุนายน กรกฎาคม สิงหาคม กันยายน ตุลาคม พฤศจิกายน ธันวาคม'.split()[now1.month] + thai_year = now1.year + 543 + time_str = now1.strftime('%H:%M:%S') + return "%d %s %d %s"%(now1.day, month_name, thai_year, time_str) # 30 ตุลาคม 2560 20:45:30
+ +def now_reign_year(): + ''' + ปีรัชกาลที่ 10 + ณ ปัจจุบัน + ''' + return now1.year - 2015 +def reign_year_to_ad(reign_year,reign): + ''' + ปีรัชกาล แปลงเป็น ค.ศ. + reign_year_to_ad(reign_year,reign) + reign_year - ปีที่ + reign - รัชกาล + ''' + if int(reign)==10: + ad = int(reign_year)+2015 + elif int(reign)==9: + ad = int(reign_year)+1945 + elif int(reign)==8: + ad = int(reign_year)+1928 + elif int(reign)==7: + ad = int(reign_year)+1924 + return ad +# BE คือ พ.ศ. +# AD คือ ค.ศ. +# AH ปีฮิจเราะห์ศักราชเป็นปีพุทธศักราช จะต้องบวกด้วย 1122 +
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/ner.html b/docs/_build/html/_modules/pythainlp/ner.html new file mode 100644 index 000000000..a3894745b --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/ner.html @@ -0,0 +1,346 @@ + + + + + + + + + + + pythainlp.ner — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.ner

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import,unicode_literals
+# NLP
+import re
+from pythainlp.tokenize import word_tokenize
+from pythainlp.tag import pos_tag
+from pythainlp.corpus import stopwords
+thaicut="newmm" # ตัวตัดคำ
+# CRF
+try:
+    import sklearn_crfsuite
+except ImportError:
+    from pythainlp.tools import install_package
+    install_package('sklearn-crfsuite')
+    import sklearn_crfsuite
+# FILE
+import glob
+import codecs
+from pythainlp.corpus import get_file,download
+
+stopwords = stopwords.words('thai')
+
+
+def isThai(chr): # เช็คว่าเป็น char ภาษาไทย
+ cVal = ord(chr)
+ if(cVal >= 3584 and cVal <= 3711):
+  return True
+ return False
+def isThaiWord(word): # เช็คว่าเป็นคำภาษาไทย
+ t=True
+ for i in word:
+  l=isThai(i)
+  if l!=True and i!='.':
+   t=False
+   break
+ return t
+
+def is_stopword(word): # เช็คว่าเป็นคำฟุ่งเฟือง
+    return word in stopwords
+def doc2features(doc, i):
+    word = doc[i][0]
+    postag = doc[i][1]
+    # Features from current word
+    features={
+        'word.word': word,
+        'word.stopword': is_stopword(word),
+        'word.isthai':isThaiWord(word),
+        'word.isspace':word.isspace(),
+        'postag':postag,
+        'word.isdigit()': word.isdigit()
+    }
+    if word.isdigit() and len(word)==5:
+        features['word.islen5']=True
+    if i > 0:
+        prevword = doc[i-1][0]
+        postag1 = doc[i-1][1]
+        features['word.prevword'] = prevword
+        features['word.previsspace']=prevword.isspace()
+        features['word.previsthai']=isThaiWord(prevword)
+        features['word.prevstopword']=is_stopword(prevword)
+        features['word.prepostag'] = postag1
+        features['word.prevwordisdigit'] = prevword.isdigit()
+    else:
+        features['BOS'] = True # Special "Beginning of Sequence" tag
+    # Features from next word
+    if i < len(doc)-1:
+        nextword = doc[i+1][0]
+        postag1 = doc[i+1][1]
+        features['word.nextword'] = nextword
+        features['word.nextisspace']=nextword.isspace()
+        features['word.nextpostag'] = postag1
+        features['word.nextisthai']=isThaiWord(nextword)
+        features['word.nextstopword']=is_stopword(nextword)
+        features['word.nextwordisdigit'] = nextword.isdigit()
+    else:
+        features['EOS'] = True # Special "End of Sequence" tag
+    return features
+
+
[docs]class thainer: + def __init__(self): + """ + Thai NER + """ + self.data_path = get_file('thainer') + if self.data_path==None: + download('thainer') + self.data_path = get_file('thainer') + self.crf=sklearn_crfsuite.CRF( + algorithm='lbfgs', + c1=0.1, + c2=0.1, + max_iterations=500, + all_possible_transitions=True, + model_filename=self.data_path) +
[docs] def get_ner(self,text,postag=True): + """ + Get NER from Thai NER. + + :param string text: thai text + :param boolean postag: get postag (True) or get not postag (False) + + :return: list NER. + + **Example**:: + >>> from pythainlp.ner import thainer + >>> ner=thainer() + >>> ner.get_ner("วันที่ 15 ก.ย. 61 ทดสอบระบบเวลา 14:49 น.") + [('วันที่', 'JSBR', 'O'), (' ', 'NCMN', 'O'), ('15', 'NCNM', 'B-DATE'), (' ', 'NCMN', 'I-DATE'), ('ก.ย.', 'CMTR', 'I-DATE'), (' ', 'NCMN', 'I-DATE'), ('61', 'NCNM', 'I-DATE'), (' ', 'NCMN', 'O'), ('ทดสอบ', 'VACT', 'O'), ('ระบบ', 'NCMN', 'O'), ('เวลา', 'NCMN', 'O'), (' ', 'NCMN', 'O'), ('14', 'NCNM', 'B-TIME'), (':', 'PUNC', 'I-TIME'), ('49', 'NCNM', 'I-TIME'), (' ', 'NCMN', 'I-TIME'), ('น.', 'CMTR', 'I-TIME')] + >>> ner.get_ner("วันที่ 15 ก.ย. 61 ทดสอบระบบเวลา 14:49 น.",postag=False) + [('วันที่', 'O'), (' ', 'O'), ('15', 'B-DATE'), (' ', 'I-DATE'), ('ก.ย.', 'I-DATE'), (' ', 'I-DATE'), ('61', 'I-DATE'), (' ', 'O'), ('ทดสอบ', 'O'), ('ระบบ', 'O'), ('เวลา', 'O'), (' ', 'O'), ('14', 'B-TIME'), (':', 'I-TIME'), ('49', 'I-TIME'), (' ', 'I-TIME'), ('น.', 'I-TIME')] + """ + self.word_cut=word_tokenize(text,engine=thaicut) + self.list_word=pos_tag(self.word_cut,engine='perceptron') + self.X_test = self.extract_features([(data,self.list_word[i][1]) for i,data in enumerate(self.word_cut)]) + self.y_=self.crf.predict_single(self.X_test) + if postag: + return [(self.word_cut[i],self.list_word[i][1],data) for i,data in enumerate(self.y_)] + else: + return [(self.word_cut[i],data) for i,data in enumerate(self.y_)]
+ def extract_features(self,doc): + return [doc2features(doc, i) for i in range(len(doc))] + def get_labels(self,doc): + return [tag for (token,postag,tag) in doc] + def get_model(self): + return self.crf
+
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/number.html b/docs/_build/html/_modules/pythainlp/number.html new file mode 100644 index 000000000..d564e3df6 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/number.html @@ -0,0 +1,354 @@ + + + + + + + + + + + pythainlp.number — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.number

+# -*- coding: utf-8 -*-
+''' ระบบแปลงเลขใน 1- 10 ภาษาไทย
+fork by http://justmindthought.blogspot.com/2012/12/code-php.html
+'''
+from __future__ import absolute_import,division,print_function,unicode_literals
+from builtins import dict
+from builtins import int
+import math,six,ast
+p = [[u'ภาษาไทย', u'ตัวเลข',u'เลขไทย'],
+     [u'หนึ่ง', u'1', u'๑'],
+     [u'สอง', u'2', u'๒'],
+     [u'สาม', u'3', u'๓'],
+     [u'สี่', u'4', u'๔'],
+     [u'ห้า', u'5', u'๕'],
+     [u'หก', u'6', u'๖'],
+     [u'หก', u'7', u'๗'],
+     [u'แปด', u'8', u'๘'],
+     [u'เก้า', u'9', u'๙']]
+thaitonum = dict((x[2], x[1]) for x in p[1:])
+p1 = dict((x[0], x[1]) for x in p[1:])
+d1 = 0
+#เลขไทยสู่เลข
+
[docs]def thai_num_to_num(text): + """ + :param str text: Thai number characters such as '๑', '๒', '๓' + :return: universal numbers such as '1', '2', '3' + """ + thaitonum = dict((x[2], x[1]) for x in p[1:]) + return thaitonum[text]
+ +
[docs]def thai_num_to_text(text): + """ + :param str text: Thai number characters such as '๑', '๒', '๓' + :return: Thai numbers, spelled out in Thai + """ + thaitonum = dict((x[2], x[0]) for x in p[1:]) + return thaitonum[text]
+ +
[docs]def num_to_thai_num(text): + """ + :param text: universal numbers such as '1', '2', '3' + :return: Thai number characters such as '๑', '๒', '๓' + """ + thaitonum = dict((x[1], x[2]) for x in p[1:]) + return thaitonum[text]
+ +
[docs]def num_to_text(text): + """ + :param text: universal numbers such as '1', '2', '3' + :return: Thai numbers, spelled out in Thai + """ + thaitonum = dict((x[1], x[0]) for x in p[1:]) + return thaitonum[text]
+ +
[docs]def text_to_num(text): + """ + :param text: Thai numbers, spelled out in Thai + :return: universal numbers such as '1', '2', '3' + """ + thaitonum = dict((x[0], x[1]) for x in p[1:]) + return thaitonum[text]
+ +def text_to_thai_num(text): + """ + :param text: Thai numbers, spelled out in Thai + :return: Thai numbers such as '๑', '๒', '๓' + """ + thaitonum = dict((x[0], x[2]) for x in p[1:]) + return thaitonum[text] + +def number_format(num, places=0): + return '{:20,.2f}'.format(num) +# fork by http://justmindthought.blogspot.com/2012/12/code-php.html + +
[docs]def numtowords(amount_number): + amount_number = number_format(amount_number, 2).replace(" ","") + pt = amount_number.find(".") + number,fraction = "","" + amount_number1 = amount_number.split('.') + if (pt == False): + number = amount_number + else: + amount_number = amount_number.split('.') + number = amount_number[0] + fraction = int(amount_number1[1]) + ret = "" + number=ast.literal_eval(number.replace(",","")) + baht = readnumber(number) + if (baht != ""): + ret += baht + "บาท" + satang = readnumber(fraction) + if (satang != ""): + ret += satang + "สตางค์" + else: + ret += "ถ้วน" + return ret
+ +def readnumber(number): + """ + :param float number: a float number (with decimals) indicating a quantity + :return: a text that indicates the full amount in word form, properly ending each digit with the right term. + """ + position_call = ["แสน", "หมื่น", "พัน", "ร้อย", "สิบ", ""] + number_call = ["", "หนึ่ง", "สอง", "สาม","สี่", "ห้า", "หก", "เจ็ด", "แปด", "เก้า"] + number = number + ret = "" + if (number == 0): return ret + if (number > 1000000): + ret += readnumber(int(number / 1000000)) + "ล้าน" + number = int(math.fmod(number, 1000000)) + divider = 100000 + pos = 0 + while(number > 0): + d=int(number/divider) + if (divider == 10) and (d == 2): + ret += "ยี่" + elif (divider == 10) and (d == 1): + ret += "" + elif ((divider == 1) and (d == 1) and (ret != "")): + ret += "เอ็ด" + else: + ret += number_call[d] + if d: + ret += position_call[pos] + else: + ret += "" + number=number % divider + divider=divider / 10 + pos += 1 + return ret + +if __name__ == "__main__": + print(numtowords(4000.0)) +
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/romanization.html b/docs/_build/html/_modules/pythainlp/romanization.html new file mode 100644 index 000000000..09b66e99e --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/romanization.html @@ -0,0 +1,248 @@ + + + + + + + + + + + pythainlp.romanization — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.romanization

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import,unicode_literals
+from pythainlp.tokenize import word_tokenize
+# ถอดเสียงภาษาไทยเป็น Latin
+
[docs]def romanization(data,engine='royin'): + """ + :param str data: Thai text to be romanized + :param str engine: choose between 'royin' , 'pyicu' and 'thai2rom'. 'royin' will romanize according to the standard of Thai Royal Institute. 'pyicu' will romanize according to the Internaitonal Phonetic Alphabet. 'thai2rom' is deep learning thai romanization. + :return: English (more or less) text that spells out how the Thai text should read. + """ + word_list=word_tokenize(data) + listword=[] + i=0 + if engine=='royin': + from .royin import romanization + elif engine=='pyicu': + from .pyicu import romanization + elif engine=='thai2rom': + from pythainlp.romanization.thai2rom import thai2rom + thai=thai2rom() + return thai.romanization(data) + else: + raise Exception("error no have engine.") + while i<len(word_list): + listword.append(romanization(word_list[i])) + i+=1 + return ''.join(listword)
+
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/romanization/thai2rom.html b/docs/_build/html/_modules/pythainlp/romanization/thai2rom.html new file mode 100644 index 000000000..072aa8dd1 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/romanization/thai2rom.html @@ -0,0 +1,338 @@ + + + + + + + + + + + pythainlp.romanization.thai2rom — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.romanization.thai2rom

+# -*- coding: utf-8 -*-
+from __future__ import print_function
+
+try:
+    import numpy as np
+    import keras
+except ImportError:
+    from pythainlp.tools import install_package
+    install_package('keras')
+    install_package('numpy')
+
+from pythainlp.corpus import get_file,download
+
+from keras.models import Model, load_model
+from keras.layers import Input
+import numpy as np
+
[docs]class thai2rom: + def __init__(self): + ''' + Thai2Rom + ''' + self.batch_size = 64 + self.epochs = 100 + self.latent_dim = 256 + self.num_samples = 648241 + self.data_path = get_file('thai2rom-dataset') + if self.data_path==None: + download('thai2rom-dataset') + self.data_path = get_file('thai2rom-dataset') + self.input_texts = [] + self.target_texts = [] + self.input_characters = set() + self.target_characters = set() + with open(self.data_path, 'r', encoding='utf-8-sig') as self.f: + self.lines = self.f.read().split('\n') + for self.line in self.lines[: min(self.num_samples, len(self.lines) - 1)]: + self.input_text, self.target_text = self.line.split('\t') + if len(self.input_text)<30 and len(self.target_text)<90: + self.target_text = '\t' + self.target_text + '\n' + self.input_texts.append(self.input_text) + self.target_texts.append(self.target_text) + for self.char in self.input_text: + if self.char not in self.input_characters: + self.input_characters.add(self.char) + for self.char in self.target_text: + if self.char not in self.target_characters: + self.target_characters.add(self.char) + self.input_characters = sorted(list(self.input_characters)) + self.target_characters = sorted(list(self.target_characters)) + self.num_encoder_tokens = len(self.input_characters) + self.num_decoder_tokens = len(self.target_characters) + self.max_encoder_seq_length = max([len(self.txt) for self.txt in self.input_texts]) + self.max_decoder_seq_length = max([len(self.txt) for self.txt in self.target_texts]) + '''print('Number of samples:', len(self.input_texts)) + print('Number of unique input tokens:', self.num_encoder_tokens) + print('Number of unique output tokens:', self.num_decoder_tokens) + print('Max sequence length for inputs:', self.max_encoder_seq_length) + print('Max sequence length for outputs:', self.max_decoder_seq_length)''' + self.input_token_index = dict([(char, i) for i, char in enumerate(self.input_characters)]) + self.target_token_index = dict([(char, i) for i, char in enumerate(self.target_characters)]) + self.encoder_input_data = np.zeros((len(self.input_texts), self.max_encoder_seq_length, self.num_encoder_tokens),dtype='float32') + for i, input_text in enumerate(self.input_texts): + for t, char in enumerate(self.input_text): + self.encoder_input_data[i, t, self.input_token_index[char]] = 1. + # Restore the model and construct the encoder and decoder. + self.filemodel=get_file('thai2rom') + if self.filemodel==None: + download('thai2rom') + self.filemodel=get_file('thai2rom') + self.model = load_model(self.filemodel) + self.encoder_inputs = self.model.input[0] # input_1 + self.encoder_outputs, self.state_h_enc, self.state_c_enc = self.model.layers[2].output # lstm_1 + self.encoder_states = [self.state_h_enc, self.state_c_enc] + self.encoder_model = Model(self.encoder_inputs, self.encoder_states) + self.decoder_inputs = self.model.input[1] # input_2 + self.decoder_state_input_h = Input(shape=(self.latent_dim,), name='input_3') + self.decoder_state_input_c = Input(shape=(self.latent_dim,), name='input_4') + self.decoder_states_inputs = [self.decoder_state_input_h, self.decoder_state_input_c] + self.decoder_lstm = self.model.layers[3] + self.decoder_outputs, self.state_h_dec, self.state_c_dec = self.decoder_lstm(self.decoder_inputs, initial_state=self.decoder_states_inputs) + self.decoder_states = [self.state_h_dec, self.state_c_dec] + self.decoder_dense = self.model.layers[4] + self.decoder_outputs = self.decoder_dense(self.decoder_outputs) + self.decoder_model = Model([self.decoder_inputs] + self.decoder_states_inputs,[self.decoder_outputs] + self.decoder_states) + + self.reverse_input_char_index = dict((i, char) for char, i in self.input_token_index.items()) + self.reverse_target_char_index = dict((i, char) for char, i in self.target_token_index.items()) + def decode_sequence(self,input_seq): + self.states_value = self.encoder_model.predict(input_seq) + self.target_seq = np.zeros((1, 1, self.num_decoder_tokens)) + self.target_seq[0, 0, self.target_token_index['\t']] = 1. + self.stop_condition = False + self.decoded_sentence = '' + while not self.stop_condition: + self.output_tokens, self.h, self.c = self.decoder_model.predict([self.target_seq] + self.states_value) + self.sampled_token_index = np.argmax(self.output_tokens[0, -1, :]) + self.sampled_char = self.reverse_target_char_index[self.sampled_token_index] + self.decoded_sentence += self.sampled_char + if (self.sampled_char == '\n' or len(self.decoded_sentence) > self.max_decoder_seq_length): + self.stop_condition = True + self.target_seq = np.zeros((1, 1, self.num_decoder_tokens)) + self.target_seq[0, 0, self.sampled_token_index] = 1. + self.states_value = [self.h, self.c] + return self.decoded_sentence + def encode_input(self,name): + self.test_input = np.zeros((1, self.max_encoder_seq_length, self.num_encoder_tokens),dtype='float32') + for t, char in enumerate(name): + self.test_input[0, t, self.input_token_index[char]] = 1. + return self.test_input +
[docs] def romanization(self,text): + ''' + :param str text: Thai text to be romanized + :return: English (more or less) text that spells out how the Thai text should read. + ''' + return self.decode_sequence(self.encode_input(text))
+
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/sentiment.html b/docs/_build/html/_modules/pythainlp/sentiment.html new file mode 100644 index 000000000..54137eba9 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/sentiment.html @@ -0,0 +1,271 @@ + + + + + + + + + + + pythainlp.sentiment — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.sentiment

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import,unicode_literals,print_function
+import pythainlp
+from pythainlp.corpus import stopwords
+import os
+from pythainlp.tokenize import word_tokenize
+import dill
+
+templates_dir = os.path.join(os.path.dirname(pythainlp.__file__), 'sentiment')
+
[docs]def sentiment(text, engine='old'): + """ + :param str text: thai text + :param str engine: sentiment analysis engine (old or ulmfit) + :return: pos or neg + + **Example**:: + >>> from pythainlp.sentiment import sentiment + >>> text="วันนี้อากาศดีจัง" + >>> sentiment(text) + 'pos' + >>> sentiment(text,'ulmfit') + 'pos' + >>> text="วันนี้อารมณ์เสียมาก" + >>> sentiment(text) + 'neg' + >>> sentiment(text,'ulmfit') + 'neg' + """ + if engine=='old': + with open(os.path.join(templates_dir, 'vocabulary.data'), 'rb') as in_strm: + vocabulary = dill.load(in_strm) + with open(os.path.join(templates_dir, 'sentiment.data'), 'rb') as in_strm: + classifier = dill.load(in_strm) + text=set(word_tokenize(text))-set(stopwords.words('thai')) + featurized_test_sentence = {i:(i in text) for i in vocabulary} + return classifier.classify(featurized_test_sentence) + elif engine=='ulmfit': + from pythainlp.sentiment import ulmfit_sent + tag=ulmfit_sent.get_sentiment(text) + sa="" + if tag==0: + sa="neg" + else: + sa="pos" + return sa + else: + raise Exception("error no have engine.")
+if __name__ == '__main__': + d="เสียใจแย่มากเลย" + print(sentiment(d)) +
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/soundex.html b/docs/_build/html/_modules/pythainlp/soundex.html new file mode 100644 index 000000000..a426ec9dc --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/soundex.html @@ -0,0 +1,310 @@ + + + + + + + + + + + pythainlp.soundex — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.soundex

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import,division,unicode_literals,print_function
+from builtins import *
+'''
+Thai soundex
+
+โค้ดพัฒนาโดย คุณ Korakot Chaovavanich (จาก https://gist.github.com/korakot/0b772e09340cac2f493868da035597e8)
+'''
+import re
+
[docs]def LK82(s): + ''' + LK82 - It's a thai soundex rule. + + :param str s: thai word + :return: LK82 soundex + ''' + t1 = str.maketrans("กขฃคฅฆงจฉชฌซศษสญยฎดฏตณนฐฑฒถทธบปผพภฝฟมรลฬฤฦวหฮอ","กกกกกกงจชชชซซซซยยดดตตนนททททททบปพพพฟฟมรรรรรวหหอ") + t2 = str.maketrans("กขฃคฅฆงจฉชซฌฎฏฐฑฒดตถทธศษสญณนรลฬฤฦบปพฟภผฝมำยวไใหฮาๅึืเแโุูอ","1111112333333333333333333444444445555555667777889AAABCDEEF") + res = [] + s = re.sub("[่-๋]", "", s) # 4.ลบวรรณยุกต์ + s = re.sub('จน์|มณ์|ณฑ์|ทร์|ตร์|[ก-ฮ]์|[ก-ฮ][ะ-ู]์', "", s) # 4.ลบตัวการันต์ + s = re.sub("[็ํฺๆฯ]", "", s) # 5.ทิ้งไม้ไต่คู่ ฯลฯ + # 6.เข้ารหัสตัวแรก + if 'ก'<=s[0]<='ฮ': + res.append(s[0].translate(t1)) + s = s[1:] + else: + res.append(s[1].translate(t1)) + res.append(s[0].translate(t2)) + s = s[2:] + # เข้ารหัสตัวที่เหลือ + i_v = None # ตำแหน่งตัวคั่นล่าสุด (สระ) + for i,c in enumerate(s): + if c in "ะัิี": # 7. ตัวคั่นเฉยๆ + i_v = i + res.append('') + elif c in "าๅึืู": # 8.คั่นและใส่ + i_v = i + res.append(c.translate(t2)) + elif c == 'ุ': # 9.สระอุ + i_v = i + if i==0 or (s[i-1] not in "ตธ"): + res.append(c.translate(t2)) + else: + res.append('') + elif c in 'หอ': + if i+1<len(s) and (s[i+1] in "ึืุู"): + res.append(c.translate(t2)) + elif c in 'รวยฤฦ': + if i_v == i-1 or (i+1<len(s) and (s[i+1] in "ึืุู")): + res.append(c.translate(t2)) + else: + res.append(c.translate(t2)) # 12. + # 13. เอาตัวซ้ำออก + res2 = [res[0]] + for i in range(1, len(res)): + if res[i] != res[i-1]: + res2.append(res[i]) + # 14. เติมศูนย์ให้ครบ ถ้าเกินก็ตัด + return ("".join(res2)+"0000")[:5]
+
[docs]def Udom83(s): + ''' + Udom83 - It's a thai soundex rule. + + :param str s: thai word + :return: LK82 soundex + ''' + tu1 = str.maketrans("กขฃคฅฆงจฉชฌซศษสฎดฏตฐฑฒถทธณนบปผพภฝฟมญยรลฬฤฦวอหฮ" ,"กขขขขขงจชชชสสสสดดตตททททททนนบปพพพฟฟมยยรรรรรวอฮฮ") + tu2 = str.maketrans("มวำกขฃคฅฆงยญณนฎฏดตศษสบปพภผฝฟหอฮจฉชซฌฐฑฒถทธรฤลฦ","0001111112233344444445555666666777778888889999") + s = re.sub('รร([เ-ไ])', 'ัน\\1', s) # 4. + s = re.sub('รร([ก-ฮ][ก-ฮเ-ไ])', 'ั\\1', s) # 5. + s = re.sub('รร([ก-ฮ][ะ-ู่-์])','ัน\\1', s) + s = re.sub('รร', 'ัน', s) + s = re.sub('ไ([ก-ฮ]ย)', '\\1', s) # 2. + s = re.sub('[ไใ]([ก-ฮ])','\\1ย', s) + s = re.sub('ำ(ม[ะ-ู])', 'ม\\1', s) # 3. + s = re.sub('ำม', 'ม', s) + s = re.sub('ำ', 'ม', s) + s = re.sub('จน์|มณ์|ณฑ์|ทร์|ตร์|[ก-ฮ]์|[ก-ฮ][ะ-ู]์', "", s) # 6. + s = re.sub('[ะ-์]', '', s) # 7. + sd = s[0].translate(tu1) + sd += s[1:].translate(tu2) + return (sd+'000000')[:7]
+if __name__ == '__main__': + print(LK82('รถ')) + print(LK82('รส')) + print(LK82('รด')) + print(LK82('จัน')) + print(LK82('จันทร์')) +
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/spell.html b/docs/_build/html/_modules/pythainlp/spell.html new file mode 100644 index 000000000..ba2abdde9 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/spell.html @@ -0,0 +1,236 @@ + + + + + + + + + + + pythainlp.spell — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.spell

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import,unicode_literals
+
[docs]def spell(word,engine='pn'): + """ + :param str word: the word to check spelling + :param str engine: + * pn - Peter Norvig's algorithm + * hunspell - uses hunspell's algorithm, which should already exist in linux + :return: list word + """ + if engine=='pn': + from .pn import spell as spell1 + elif engine=='hunspell': + from .hunspell import spell as spell1 + return spell1(word)
+
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/summarize.html b/docs/_build/html/_modules/pythainlp/summarize.html new file mode 100644 index 000000000..ecca8c3a4 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/summarize.html @@ -0,0 +1,272 @@ + + + + + + + + + + + pythainlp.summarize — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.summarize

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import,unicode_literals
+from pythainlp.corpus import stopwords
+from string import punctuation
+from collections import defaultdict
+from pythainlp.tokenize import sent_tokenize, word_tokenize
+from heapq import nlargest
+class FrequencySummarizer:
+    def __init__(self, min_cut=0.1, max_cut=0.9):
+        self._min_cut = min_cut
+        self._max_cut = max_cut
+        self._stopwords = set(stopwords.words('thai') + list(punctuation))
+
+    def _compute_frequencies(self, word_sent):
+        freq = defaultdict(int)
+        for s in word_sent:
+            for word in s:
+                if word not in self._stopwords:
+                    freq[word] += 1
+        m = float(max(freq.values()))
+        for w in list(freq):
+            freq[w] = freq[w]/m
+            if freq[w] >= self._max_cut or freq[w] <= self._min_cut:
+                del freq[w]
+        return freq
+
+    def _rank(self, ranking, n):
+        return nlargest(n, ranking, key=ranking.get)
+
+    def summarize(self, text, n,tokenize):
+        sents = sent_tokenize(text)
+        word_sent = [word_tokenize(s,tokenize) for s in sents]
+        self._freq = self._compute_frequencies(word_sent)
+        ranking = defaultdict(int)
+        for i, sent in enumerate(word_sent):
+            for w in sent:
+                if w in self._freq:
+                    ranking[i] += self._freq[w]
+        sents_idx = self._rank(ranking,n)
+        return [sents[j] for j in sents_idx]
+
[docs]def summarize_text(text,n,engine='frequency',tokenize='newmm'): + ''' + Thai text summarize. + :param str text: thai text + :param int n: sent number + :param str engine: Thai text summarize engine. + :param str tokenize: thai word tokenize. + ''' + if engine=='frequency': + data=FrequencySummarizer().summarize(text,n,tokenize) + return data
+
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/tag.html b/docs/_build/html/_modules/pythainlp/tag.html new file mode 100644 index 000000000..a2c447ae4 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/tag.html @@ -0,0 +1,264 @@ + + + + + + + + + + + pythainlp.tag — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.tag

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import,division,print_function,unicode_literals
+import sys
+
[docs]def pos_tag(list_text,engine='unigram',corpus='orchid'): + """ + Part of Speech tagging function. + + :param list list_text: takes in a list of tokenized words (put differently, a list of string) + :param str engine: + * unigram - unigram tagger + * perceptron - perceptron tagger + * artagger - RDR POS tagger + :param str corpus: + * orchid - annotated Thai academic articles + * pud - Parallel Universal Dependencies (PUD) treebanks + :return: returns a list of labels regarding which part of speech it is + """ + if engine=='old' or engine=='unigram': + from .old import tag + elif engine=='perceptron': + from .perceptron import tag + elif engine=='artagger': + def tag(text1): + try: + from artagger import Tagger + except ImportError: + from pythainlp.tools import install_package + install_package('https://github.com/wannaphongcom/artagger/archive/master.zip') + try: + from artagger import Tagger + except ImportError: + print("Error ! using 'pip install https://github.com/wannaphongcom/artagger/archive/master.zip'") + sys.exit(0) + words = Tagger().tag(' '.join(text1)) + totag=[] + for word in words: + totag.append((word.word, word.tag)) + return totag + return tag(list_text) + return tag(list_text,corpus=corpus)
+ +def pos_tag_sents(sentences,engine='unigram',corpus='orchid'): + return [pos_tag(i,engine=engine,corpus=corpus) for i in sentences] +
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/tokenize.html b/docs/_build/html/_modules/pythainlp/tokenize.html new file mode 100644 index 000000000..11b0af8f9 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/tokenize.html @@ -0,0 +1,422 @@ + + + + + + + + + + + pythainlp.tokenize — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.tokenize

+# -*- coding: utf-8 -*-
+from __future__ import absolute_import,unicode_literals
+import nltk
+import re
+import codecs
+from six.moves import zip
+from pythainlp.corpus.thaisyllable import get_data
+from pythainlp.corpus.thaiword import get_data as get_dict
+from marisa_trie import Trie
+
+DEFAULT_DICT_TRIE = Trie(get_dict())
+
+
[docs]def word_tokenize(text, engine='newmm',whitespaces=True): + """ + :param str text: the text to be tokenized + :param str engine: the engine to tokenize text + :param bool whitespaces: True to output no whitespace, a common mark of sentence or end of phrase in Thai. + :Parameters for engine: + * newmm - Maximum Matching algorithm + TCC + * icu - IBM ICU + * longest-matching - Longest matching + * mm - Maximum Matching algorithm + * pylexto - LexTo + * deepcut - Deep Neural Network + * wordcutpy - wordcutpy (https://github.com/veer66/wordcutpy) + :return: A list of words, tokenized from a text + + **Example**:: + + from pythainlp.tokenize import word_tokenize + text='ผมรักคุณนะครับโอเคบ่พวกเราเป็นคนไทยรักภาษาไทยภาษาบ้านเกิด' + a=word_tokenize(text,engine='icu') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอ', 'เค', 'บ่', 'พวก', 'เรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้าน', 'เกิด'] + b=word_tokenize(text,engine='dict') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + c=word_tokenize(text,engine='mm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + d=word_tokenize(text,engine='pylexto') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + e=word_tokenize(text,engine='newmm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + g=word_tokenize(text,engine='wordcutpy') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้านเกิด'] + """ + if engine=='icu': + from .pyicu import segment + elif engine=='multi_cut' or engine=='mm': + from .multi_cut import segment + elif engine=='newmm' or engine=='onecut': + from .newmm import mmcut as segment + elif engine=='longest-matching': + from .longest import segment + elif engine=='pylexto': + from .pylexto import segment + elif engine=='deepcut': + from .deepcut import segment + elif engine=='wordcutpy': + from .wordcutpy import segment + else: + raise Exception("error no have engine.") + if whitespaces==False: + return [i.strip(' ') for i in segment(text) if i.strip(' ')!=''] + return segment(text)
+
[docs]def dict_word_tokenize(text, custom_dict_trie, engine='newmm'): + ''' + :meth:`dict_word_tokenize` tokenizes word based on the dictionary you provide. The format has to be in trie data structure. + + :param str text: the text to be tokenized + :param dict custom_dict_trie: คือ trie ที่สร้างจาก create_custom_dict_trie + :param str engine: choose between different options of engine to token (newmm, wordcutpy, mm, longest-matching) + :return: A list of words, tokenized from a text. + **Example**:: + >>> from pythainlp.tokenize import dict_word_tokenize,create_custom_dict_trie + >>> listword=['แมว',"ดี"] + >>> data_dict=create_custom_dict_trie(listword) + >>> dict_word_tokenize("แมวดีดีแมว",data_dict) + ['แมว', 'ดี', 'ดี', 'แมว'] + ''' + if engine=="newmm" or engine=="onecut": + from .newmm import mmcut as segment + elif engine=="mm" or engine=="multi_cut": + from .multi_cut import segment + elif engine=='longest-matching': + from .longest import segment + elif engine=='wordcutpy': + from .wordcutpy import segment + return segment(text, custom_dict_trie.keys()) + else: + raise Exception("error no have engine.") + return segment(text, custom_dict_trie)
+
[docs]def sent_tokenize(text,engine='whitespace+newline'): + ''' + This function does not yet automatically recognize when a sentence actually ends. Rather it helps split text where white space and a new line is found. + + :param str text: the text to be tokenized + :param str engine: choose between 'whitespace' or 'whitespace+newline' + + :return: a list of text, split by whitespace or new line. + ''' + if engine=='whitespace': + data=nltk.tokenize.WhitespaceTokenizer().tokenize(text) + elif engine=='whitespace+newline': + data=re.sub(r'\n+|\s+','|',text,re.U).split('|') + return data
+ +
[docs]def subword_tokenize(text, engine='tcc'): + """ + :param str text: text to be tokenized + :param str engine: choosing 'tcc' uses the Thai Character Cluster rule to segment words into the smallest unique units. + :return: a list of tokenized strings. + """ + if engine == 'tcc': + from .tcc import tcc + return tcc(text)
+ +
[docs]def isthai(text,check_all=False): + """ + :param str text: input string or list of strings + :param bool check_all: checks all character or not + + :return: A dictionary with the first value as proportional of text that is Thai, and the second value being a tuple of all characters, along with true or false. + """ + listext=list(text) + i=0 + num_isthai=0 + if check_all==True: + listthai=[] + while i<len(listext): + cVal = ord(listext[i]) + if(cVal >= 3584 and cVal <= 3711): + num_isthai+=1 + if check_all==True: + listthai.append(True) + else: + if check_all==True: + listthai.append(False) + i+=1 + thai=(num_isthai/len(listext))*100 + if check_all==True: + dictthai=tuple(zip(listext,listthai)) + data= {'thai':thai,'check_all':dictthai} + else: + data= {'thai':thai} + return data
+ +def syllable_tokenize(text): + """ + :param str text: input string to be tokenized + + :return: returns list of strings of syllables + """ + text1=word_tokenize(text) + data=[] + trie = create_custom_dict_trie(custom_dict_source=get_data()) + if len(text1)>1: + i=0 + while i<len(text1): + data.extend(dict_word_tokenize(text=text1[i], custom_dict_trie=trie)) + i+=1 + else: + data=dict_word_tokenize(text=text, custom_dict_trie=trie) + return data + +
[docs]def create_custom_dict_trie(custom_dict_source): + """The function is used to create a custom dict trie which will be used for word_tokenize() function. For more information on the trie data structure, see: https://marisa-trie.readthedocs.io/en/latest/index.html + + :param string/list custom_dict_source: a list of vocaburaries or a path to source file + + :return: A trie created from custom dict input + """ + + if type(custom_dict_source) is str: + # Receive a file path of the custom dict to read + with codecs.open(custom_dict_source, 'r',encoding='utf8') as f: + _vocabs = f.read().splitlines() + return Trie(_vocabs) + elif isinstance(custom_dict_source, (list, tuple, set)): + # Received a sequence type object of vocabs + return Trie(custom_dict_source) + else: + raise TypeError( + 'Type of custom_dict_source must be either str (path to source file) or collections' + )
+ +class Tokenizer: + def __init__(self, custom_dict=None): + """ + Initialize tokenizer object + + :param str custom_dict: a file path or a list of vocaburaies to be used to create a trie (default - original lexitron) + + :return: trie_dict - a dictionary in the form of trie data for tokenizing engines + """ + if custom_dict: + if type(custom_dict) is list: + self.trie_dict = Trie(custom_dict) + elif type(custom_dict) is str: + with codecs.open(custom_dict, 'r',encoding='utf8') as f: + vocabs = f.read().splitlines() + self.trie_dict = Trie(vocabs) + else: + self.trie_dict = Trie(get_dict()) + + def word_tokenize(self, text, engine='newmm'): + from .newmm import mmcut as segment + return segment(text, self.trie_dict) + +
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/ulmfit/utils.html b/docs/_build/html/_modules/pythainlp/ulmfit/utils.html new file mode 100644 index 000000000..996217c84 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/ulmfit/utils.html @@ -0,0 +1,479 @@ + + + + + + + + + + + pythainlp.ulmfit.utils — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +

Source code for pythainlp.ulmfit.utils

+# -*- coding: utf-8 -*-
+'''
+Code by https://github.com/cstorm125/thai2vec/tree/master/notebook
+'''
+from __future__ import absolute_import,unicode_literals
+import os
+import sys
+import re
+import torch
+
+#numpy and fastai
+try:
+    import numpy as np
+    from fastai.text import *
+    import dill as pickle
+except ImportError:
+    from pythainlp.tools import install_package
+    install_package('fastai')
+    install_package('numpy')
+    try:
+        import numpy as np
+        from fastai.text import *
+        import dill as pickle
+    except ImportError:
+        print("Error installing using 'pip install fastai numpy dill'")
+        sys.exit(0)
+
+#import torch
+try:
+    import torch
+except ImportError:
+    print('PyTorch required. See https://pytorch.org/.')
+
+from pythainlp.tokenize import word_tokenize
+from pythainlp.corpus import get_file
+from pythainlp.corpus import download
+MODEL_NAME = 'thwiki_model2'
+ITOS_NAME = 'itos'
+
+#paralellized thai tokenizer with some text cleaning
+
[docs]class ThaiTokenizer(): + def __init__(self, engine='newmm'): + """ + :parameters for tokenization engine: + * newmm - Maximum Matching algorithm + TCC + * icu - IBM ICU + * longest-matching - Longest matching + * mm - Maximum Matching algorithm + * pylexto - LexTo + * deepcut - Deep Neural Network + """ + self.engine = engine + self.re_br = re.compile(r'<\s*br\s*/?>', re.IGNORECASE) + self.re_rep = re.compile(r'(\S)(\1{3,})') + +
[docs] def sub_br(self,text): + """ + :meth:`sub_br` replace `<br>` tags with `\n` + :param str text: text to process + :return: procssed text + """ + return self.re_br.sub("\n", text)
+ +
[docs] def tokenize(self,text): + """ + :meth: tokenize text with selected engine + :param str text: text to tokenize + :return: tokenized text + """ + return [t for t in word_tokenize(self.sub_br(text),engine=self.engine)]
+ +
[docs] @staticmethod + def replace_rep(text): + ''' + :meth:`replace_rep` replace 3 or above repetitive characters with `tkrep` + :param str text: text to process + :return: processed text where repetitions are replaced by `tkrep` followed by number of repetitions + **Example**:: + >>> from pythainlp.ulmfit.utils import ThaiTokenizer + >>> tt = ThaiTokenizer() + >>> tt.replace_rep('คือดียยยยยย') + คือดีtkrep6ย + ''' + TK_REP = 'tkrep' + c,cc = text.groups() + return f'{TK_REP}{len(cc)+1}{c}'
+ +
[docs] def proc_text(self, text): + """ + :meth: `proc_text` procss and tokenize text removing repetitions, special characters, double spaces + :param str text: text to process + :return: processed and tokenized text + """ + s = self.re_rep.sub(ThaiTokenizer.replace_rep, text) + s = re.sub(r'([/#])', r' \1 ', s) + #remvoe double space + s = re.sub(' {2,}', ' ', s) + return self.tokenize(s)
+ +
[docs] @staticmethod + def proc_all(ss): + """ + :meth: `proc_all` runs `proc_text` for multiple sentences + :param str text: text to process + :return: processed and tokenized text + """ + tok = ThaiTokenizer() + return [tok.proc_text(s) for s in ss]
+ +
[docs] @staticmethod + def proc_all_mp(ss): + """ + :meth: `proc_all` runs `proc_text` for multiple sentences using multiple cpus + :param str text: text to process + :return: processed and tokenized text + """ + ncpus = num_cpus()//2 + with ProcessPoolExecutor(ncpus) as e: + return sum(e.map(ThaiTokenizer.proc_all, ss), [])
+ +#ulmfit helper functions +BOS = 'xbos' # beginning-of-sentence tag +
[docs]def get_texts(df): + """ + :meth: `get_texts` get tuple of tokenized texts and labels + :param pandas.DataFrame df: `pandas.DataFrame` with `label` as first column and `text` as second column + :return: + * tok - lists of tokenized texts with beginning-of-sentence tag `xbos` as first element of each list + * labels - list of labels + """ + labels = df.iloc[:,0].values.astype(np.int64) + texts = BOS+df.iloc[:,1].astype(str).apply(lambda x: x.rstrip()) + tok = ThaiTokenizer().proc_all_mp(partition_by_cores(texts)) + return(tok, list(labels))
+ +
[docs]def get_all(df): + """ + :meth: `get_all` iterate `get_texts` for all the entire `pandas.DataFrame` + :param pandas.DataFrame df: `pandas.DataFrame` with `label` as first column and `text` as second column + :return: + * tok - lists of tokenized texts with beginning-of-sentence tag `xbos` as first element of each list + * labels - list of labels + """ + tok, labels = [], [] + for i, r in enumerate(df): + tok_, labels_ = get_texts(r) + tok += tok_; + labels += labels_ + return(tok, labels)
+ +
[docs]def numericalizer(df, itos=None, max_vocab = 60000, min_freq = 2, pad_tok = '_pad_', unk_tok = '_unk_'): + """ + :meth: `numericalize` numericalize tokenized texts for: + * tokens with word frequency more than `min_freq` + * at maximum vocab size of `max_vocab` + * add unknown token `_unk_` and padding token `_pad_` in first and second position + * use integer-to-string list `itos` if avaiable e.g. ['_unk_', '_pad_','first_word','second_word',...] + :param pandas.DataFrame df: `pandas.DataFrame` with `label` as first column and `text` as second column + :param list itos: integer-to-string list + :param int max_vocab: maximum number of vocabulary (default 60000) + :param int min_freq: minimum word frequency to be included (default 2) + :param str pad_tok: padding token + :param str unk_token: unknown token + :return: + * lm - `numpy.array` of numericalized texts + * tok - lists of tokenized texts with beginning-of-sentence tag `xbos` as first element of each list + * labels - list of labels + * itos - integer-to-string list e.g. ['_unk_', '_pad_','first_word','second_word',...] + * stoi - string-to-integer dict e.g. {'_unk_':0, '_pad_':1,'first_word':2,'second_word':3,...} + * freq - `collections.Counter` for word frequency + """ + tok, labels = get_all(df) + freq = Counter(p for o in tok for p in o) + if itos is None: + itos = [o for o,c in freq.most_common(max_vocab) if c>min_freq] + itos.insert(0, pad_tok) + itos.insert(0, unk_tok) + stoi = collections.defaultdict(lambda:0, {v:k for k,v in enumerate(itos)}) + lm = np.array([[stoi[o] for o in p] for p in tok]) + return(lm,tok,labels,itos,stoi,freq)
+ +
[docs]def merge_wgts(em_sz, wgts, itos_pre, itos_cls): + """ + :param pandas.DataFrame df: `pandas.DataFrame` with `label` as first column and `text` as second column + :param int em_sz: size of embedding vectors (pretrained model is at 300) + :param wgts: saved pyTorch weights of pretrained model + :param list itos_pre: integer-to-string list of pretrained model + :param list itos_cls: integer-to-string list of current dataset + :return: merged weights of the model for current dataset + """ + vocab_size = len(itos_cls) + enc_wgts = to_np(wgts['0.encoder.weight']) + #average weight of encoding + row_m = enc_wgts.mean(0) + stoi_pre = collections.defaultdict(lambda:-1, {v:k for k,v in enumerate(itos_pre)}) + #new embedding based on classification dataset + new_w = np.zeros((vocab_size, em_sz), dtype=np.float32) + for i,w in enumerate(itos_cls): + r = stoi_pre[w] + #use pretrianed embedding if present; else use the average + new_w[i] = enc_wgts[r] if r>=0 else row_m + wgts['0.encoder.weight'] = T(new_w) + wgts['0.encoder_with_dropout.embed.weight'] = T(np.copy(new_w)) + wgts['1.decoder.weight'] = T(np.copy(new_w)) + return(wgts)
+ +#feature extractor +
[docs]def document_vector(ss, m, stoi,tok_engine='newmm'): + """ + :meth: `document_vector` get document vector using pretrained ULMFit model + :param str ss: sentence to extract embeddings + :param m: pyTorch model + :param dict stoi: string-to-integer dict e.g. {'_unk_':0, '_pad_':1,'first_word':2,'second_word':3,...} + :param str tok_engine: tokenization engine (recommend using `newmm` if you are using pretrained ULMFit model) + :return: `numpy.array` of document vector sized 300 + """ + s = word_tokenize(ss) + t = LongTensor([stoi[i] for i in s]).view(-1,1).cuda() + t = Variable(t,volatile=False) + m.reset() + pred,*_ = m[0](t) + #get average of last lstm layer along bptt + res = to_np(torch.mean(pred[-1],0).view(-1)) + return(res)
+ +class SaveFeatures(): + features=None + def __init__(self, m): self.hook = m.register_forward_hook(self.hook_fn) + def hook_fn(self, module, input, output): self.features = output + def remove(self): self.hook.remove() + +#Download pretrained models +def get_path(fname): + path = get_file(fname) + if path==None: + download(fname) + path = get_file(fname) + return(path) + +def load_pretrained_model(): + path = get_path(MODEL_NAME) + wgts = torch.load(path, map_location=lambda storage, loc: storage) + return(wgts) + +def load_pretrained_itos(): + path = get_path(ITOS_NAME) + itos = pickle.load(open(path,'rb')) + return(itos) + +
[docs]def about(): + return ''' + thai2vec + State-of-the-Art Language Modeling, Text Feature Extraction and Text Classification in Thai Language. + Created as part of pyThaiNLP with ULMFit implementation from fast.ai + + Development : Charin Polpanumas + GitHub : https://github.com/cstorm125/thai2vec + '''
+
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_modules/pythainlp/word_vector/thai2vec.html b/docs/_build/html/_modules/pythainlp/word_vector/thai2vec.html new file mode 100644 index 000000000..8ed5d3ca8 --- /dev/null +++ b/docs/_build/html/_modules/pythainlp/word_vector/thai2vec.html @@ -0,0 +1,293 @@ + + + + + + + + + + + pythainlp.word_vector.thai2vec — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ +
    + +
  • Docs »
  • + +
  • Module code »
  • + +
  • pythainlp.word_vector.thai2vec
  • + + +
  • + +
  • + +
+ + +
+
+
+
+ +

Source code for pythainlp.word_vector.thai2vec

+# -*- coding: utf-8 -*-
+'''
+Code by https://github.com/cstorm125/thai2vec/blob/master/notebooks/examples.ipynb
+'''
+from __future__ import absolute_import,unicode_literals
+import six
+import sys
+if six.PY2:
+	print("Thai sentiment in pythainlp. Not support python 2.7")
+	sys.exit(0)
+try:
+	from gensim.models import KeyedVectors
+	import numpy as np
+except ImportError:
+	from pythainlp.tools import install_package
+	install_package('gensim')
+	install_package('numpy')
+	try:
+		from gensim.models import KeyedVectors
+		import numpy as np
+	except ImportError:
+		print("Error ! using 'pip install gensim numpy'")
+		sys.exit(0)
+from pythainlp.tokenize import word_tokenize
+from pythainlp.corpus import get_file
+from pythainlp.corpus import download as download_data
+import os
+
+def download():
+	path = get_file('thai2vec02')
+	if path==None:
+		download_data('thai2vec02')
+		path = get_file('thai2vec02')
+	return path
+
[docs]def get_model(): + ''' + :return: Downloads the `gensim` model.''' + return KeyedVectors.load_word2vec_format(download(),binary=False)
+
[docs]def most_similar_cosmul(positive,negative): + ''' + การใช้งาน + input list + ''' + return get_model().most_similar_cosmul(positive=positive, negative=negative)
+
[docs]def doesnt_match(listdata): + return get_model().doesnt_match(listdata)
+
[docs]def similarity(word1,word2): + ''' + :param str word1: first word + :param str word2: second word + :return: the cosine similarity between the two word vectors + ''' + return get_model().similarity(word1,word2)
+
[docs]def sentence_vectorizer(ss,dim=300,use_mean=False): + s = word_tokenize(ss) + vec = np.zeros((1,dim)) + for word in s: + if word in get_model().wv.index2word: + vec+= get_model().wv.word_vec(word) + else: pass + if use_mean: vec /= len(s) + return(vec)
+ +
[docs]def about(): + return ''' + thai2vec + State-of-the-Art Language Modeling, Text Feature Extraction and Text Classification in Thai Language. + Created as part of pyThaiNLP with ULMFit implementation from fast.ai + + Development : Charin Polpanumas + GitHub : https://github.com/cstorm125/thai2vec + '''
+
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/_sources/api/change.rst.txt b/docs/_build/html/_sources/api/change.rst.txt new file mode 100644 index 000000000..e5b770df2 --- /dev/null +++ b/docs/_build/html/_sources/api/change.rst.txt @@ -0,0 +1,11 @@ +.. currentmodule:: pythainlp.change + +pythainlp.change +==================================== +The :class:`change` is fix incorrect input language correction. + +Modules +------- + +.. autofunction:: texttothai +.. autofunction:: texttoeng \ No newline at end of file diff --git a/docs/_build/html/_sources/api/collation.rst.txt b/docs/_build/html/_sources/api/collation.rst.txt new file mode 100644 index 000000000..a889fbea1 --- /dev/null +++ b/docs/_build/html/_sources/api/collation.rst.txt @@ -0,0 +1,8 @@ +.. currentmodule:: pythainlp.collation + + +pythainlp.collation +===================================== +The :class:`pythainlp.collation` contains a function that sorts Thai text alphabetically + +.. autofunction:: collation diff --git a/docs/_build/html/_sources/api/date.rst.txt b/docs/_build/html/_sources/api/date.rst.txt new file mode 100644 index 000000000..ec5c055d0 --- /dev/null +++ b/docs/_build/html/_sources/api/date.rst.txt @@ -0,0 +1,8 @@ +.. currentmodule:: pythainlp.date + + +pythainlp.date +===================================== +The :class:`pythainlp.date` helps output dates, as spelled out in Thai. + +.. autofunction:: now diff --git a/docs/_build/html/_sources/api/ner.rst.txt b/docs/_build/html/_sources/api/ner.rst.txt new file mode 100644 index 000000000..8810ad786 --- /dev/null +++ b/docs/_build/html/_sources/api/ner.rst.txt @@ -0,0 +1,12 @@ +.. currentmodule:: pythainlp.ner + +pythainlp.ner +==================================== +The :class:`pythainlp.ner` is named entity recognition for thai. + + +Modules +------- + +.. autoclass:: thainer + :members: get_ner diff --git a/docs/_build/html/_sources/api/number.rst.txt b/docs/_build/html/_sources/api/number.rst.txt new file mode 100644 index 000000000..459cd2b1f --- /dev/null +++ b/docs/_build/html/_sources/api/number.rst.txt @@ -0,0 +1,15 @@ +.. currentmodule:: pythainlp.number + +pythainlp.number +===================================== +The :class:`pythainlp.number` contains functions for processing thai numbers and thai words that refers to numbers. + +.. autofunction:: thai_num_to_num +.. autofunction:: thai_num_to_text +.. autofunction:: num_to_thai_num +.. autofunction:: num_to_text +.. autofunction:: text_to_num +.. autofunction:: numtowords + + + diff --git a/docs/_build/html/_sources/api/romanization.rst.txt b/docs/_build/html/_sources/api/romanization.rst.txt new file mode 100644 index 000000000..74e45f9b2 --- /dev/null +++ b/docs/_build/html/_sources/api/romanization.rst.txt @@ -0,0 +1,10 @@ +.. currentmodule:: pythainlp.romanization + +pythainlp.romanization +==================================== +The :class:`romanization.romanization` turns thai text into a romanized one (put simply, spelled with English). + +.. autofunction:: romanization +.. currentmodule:: pythainlp.romanization.thai2rom +.. autoclass:: thai2rom + :members: romanization diff --git a/docs/_build/html/_sources/api/sentiment.rst.txt b/docs/_build/html/_sources/api/sentiment.rst.txt new file mode 100644 index 000000000..10d7ce603 --- /dev/null +++ b/docs/_build/html/_sources/api/sentiment.rst.txt @@ -0,0 +1,7 @@ +.. currentmodule:: pythainlp.sentiment + +pythainlp.sentiment +==================================== +The :class:`romanization.sentiment` is sentiment analysis. + +.. autofunction:: sentiment \ No newline at end of file diff --git a/docs/_build/html/_sources/api/soundex.rst.txt b/docs/_build/html/_sources/api/soundex.rst.txt new file mode 100644 index 000000000..c61d7266f --- /dev/null +++ b/docs/_build/html/_sources/api/soundex.rst.txt @@ -0,0 +1,8 @@ +.. currentmodule:: pythainlp.soundex + +pythainlp.soundex +==================================== +The :class:`pythainlp.soundex` is soundex for thai. + +.. autofunction:: LK82 +.. autofunction:: Udom83 \ No newline at end of file diff --git a/docs/_build/html/_sources/api/spell.rst.txt b/docs/_build/html/_sources/api/spell.rst.txt new file mode 100644 index 000000000..50b379cc6 --- /dev/null +++ b/docs/_build/html/_sources/api/spell.rst.txt @@ -0,0 +1,7 @@ +.. currentmodule:: pythainlp.spell + +pythainlp.spell +===================================== +The :class:`pythainlp.spell` finds the closest correctly spelled word to the given text. + +.. autofunction:: spell diff --git a/docs/_build/html/_sources/api/summarize.rst.txt b/docs/_build/html/_sources/api/summarize.rst.txt new file mode 100644 index 000000000..d7df3988a --- /dev/null +++ b/docs/_build/html/_sources/api/summarize.rst.txt @@ -0,0 +1,10 @@ +.. currentmodule:: pythainlp.summarize + +pythainlp.summarize +==================================== +The :class:`summarize` is thai text summarize. + +Modules +------- + +.. autofunction:: summarize_text diff --git a/docs/_build/html/_sources/api/tag.rst.txt b/docs/_build/html/_sources/api/tag.rst.txt new file mode 100644 index 000000000..69315fc36 --- /dev/null +++ b/docs/_build/html/_sources/api/tag.rst.txt @@ -0,0 +1,9 @@ +.. currentmodule:: pythainlp.tag + + +pythainlp.tag +===================================== +The :class:`pythainlp.tag` contains functions that are used to tag different parts of a text. + +.. autofunction:: pos_tag + diff --git a/docs/_build/html/_sources/api/tokenizer.rst.txt b/docs/_build/html/_sources/api/tokenizer.rst.txt new file mode 100644 index 000000000..c25af5faa --- /dev/null +++ b/docs/_build/html/_sources/api/tokenizer.rst.txt @@ -0,0 +1,15 @@ +.. currentmodule:: pythainlp.tokenize +.. _tokenize-doc: + + + +pythainlp.tokenize +===================================== +The :class:`pythainlp.tokenize` contains multiple functions for tokenizing a chunk of Thai text into desirable units. + +.. autofunction:: word_tokenize +.. autofunction:: dict_word_tokenize +.. autofunction:: subword_tokenize +.. autofunction:: sent_tokenize +.. autofunction:: isthai +.. autofunction:: create_custom_dict_trie diff --git a/docs/_build/html/_sources/api/ulmfit.rst.txt b/docs/_build/html/_sources/api/ulmfit.rst.txt new file mode 100644 index 000000000..400a23899 --- /dev/null +++ b/docs/_build/html/_sources/api/ulmfit.rst.txt @@ -0,0 +1,20 @@ +.. currentmodule:: pythainlp.ulmfit.utils + + + +pythainlp.ulmfit +==================================== +The :class:`ulmfit.utils` is utils for ULMFit model. + +Modules +------- + + +.. autofunction:: get_texts +.. autofunction:: get_all +.. autofunction:: numericalizer +.. autofunction:: merge_wgts +.. autofunction:: document_vector +.. autofunction:: about +.. autoclass:: ThaiTokenizer + :members: sub_br,tokenize,replace_rep,proc_text,proc_all,proc_all_mp diff --git a/docs/_build/html/_sources/api/word_vector.rst.txt b/docs/_build/html/_sources/api/word_vector.rst.txt new file mode 100644 index 000000000..bfa44a43c --- /dev/null +++ b/docs/_build/html/_sources/api/word_vector.rst.txt @@ -0,0 +1,21 @@ +.. currentmodule:: pythainlp.word_vector.thai2vec + + + +pythainlp.word_vector +==================================== +The :class:`word_vector.thai2vec` contains functions that makes use of a pre-trained vector public data. + +Dependencies +------------ +Installation of `numpy` and `gensim` is required. + +Modules +------- + +.. autofunction:: get_model +.. autofunction:: most_similar_cosmul +.. autofunction:: doesnt_match +.. autofunction:: similarity +.. autofunction:: sentence_vectorizer +.. autofunction:: about diff --git a/docs/_build/html/_sources/index.rst.txt b/docs/_build/html/_sources/index.rst.txt new file mode 100644 index 000000000..c5e2d89cc --- /dev/null +++ b/docs/_build/html/_sources/index.rst.txt @@ -0,0 +1,36 @@ +.. pythainlp documentation master file, created by + sphinx-quickstart on Sat Jun 23 15:23:30 2018. + You can adapt this file completely to your liking, but it should at least + contain the root `toctree` directive. + +PyThaiNLP documentation +===================================== +PyThaiNLP is a Python library for natural language processing (NLP) of Thai language. + + +.. toctree:: + :glob: + :maxdepth: 1 + :caption: Notes + + notes/* + +.. toctree:: + :glob: + :maxdepth: 1 + :caption: Package reference: + + api/* + + + +Indices and tables +================== + +* :ref:`genindex` +* :ref:`modindex` +* :ref:`search` + +Apache Software License 2.0 + +Maintained by the PyThaiNLP team. diff --git a/docs/_build/html/_sources/notes/getting_started.rst.txt b/docs/_build/html/_sources/notes/getting_started.rst.txt new file mode 100644 index 000000000..2fc1285c6 --- /dev/null +++ b/docs/_build/html/_sources/notes/getting_started.rst.txt @@ -0,0 +1,27 @@ +Getting Started +===================================== +PyThaiNLP is a Python library for natural language processing (NLP) of Thai language. With this package, you can NLP tasks such as classify texts and tokenize words. + +**Tokenization Example**:: + + from pythainlp.tokenize import word_tokenize + text='ผมรักคุณนะครับโอเคบ่พวกเราเป็นคนไทยรักภาษาไทยภาษาบ้านเกิด' + a=word_tokenize(text,engine='icu') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอ', 'เค', 'บ่', 'พวก', 'เรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้าน', 'เกิด'] + b=word_tokenize(text,engine='dict') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + c=word_tokenize(text,engine='mm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + d=word_tokenize(text,engine='pylexto') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + e=word_tokenize(text,engine='newmm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + g=word_tokenize(text,engine='wordcutpy') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้านเกิ + +Thai has historically faced a lot of NLP challenges. A quick list of them include as follows: + +#. **Start-end of sentence marking** - This is arguably the biggest problem for the field of Thai NLP. The lack of end of sentence marking (EOS) makes it hard for researchers to create training sets, the basis of most research in this field. The root of the problem is two-pronged. In terms of writing system, Thai uses space to indicate both commas and periods. No letter indicates an end of a sentence. In terms of language use, Thais have a habit of starting their sentences with connector terms such as 'because', 'but', 'following', etc, making it often hard even for natives to decide where the end of sentence should be when translating. + +#. **Word segmentation** - Thai does not use space and word segmentation is not easy. It boils down to understanding the context and ruling out words that do not make sense. This is a similar issue that other Asian languages such as Japanese and Chinese face in different degrees. For languages with space, a similar but less extreme problem would be multi-word expressions, like the French word for potato — 'pomme de terre'. In Thai, the best known example is "ตา-กลม" and "ตาก-ลม". As of recent, new techniques that capture words, subwords, and letters in vectors seem poised to overcome to issue. + +**Example for classification** + +https://colab.research.google.com/drive/1cnJ6O3b1jwaHwvsMWW3oQw7f8X2Ka7Sp + + + diff --git a/docs/_build/html/_sources/notes/installation.rst.txt b/docs/_build/html/_sources/notes/installation.rst.txt new file mode 100644 index 000000000..49b251b65 --- /dev/null +++ b/docs/_build/html/_sources/notes/installation.rst.txt @@ -0,0 +1,11 @@ +Installation +===================================== +For stable version, try:: + + pip install pythainlp + +For developer version, try:: + + pip install https://github.com/PyThaiNLP/pythainlp/archive/dev.zip + +Note for Windows: marisa-trie wheels can be obtained from https://www.lfd.uci.edu/~gohlke/pythonlibs/#marisa-trie , then install it with pip, for example: `pip install marisa_trie‑0.7.5‑cp36‑cp36m‑win32.whl` \ No newline at end of file diff --git a/docs/_build/html/_static/ajax-loader.gif b/docs/_build/html/_static/ajax-loader.gif new file mode 100644 index 000000000..61faf8cab Binary files /dev/null and b/docs/_build/html/_static/ajax-loader.gif differ diff --git a/docs/_build/html/_static/basic.css b/docs/_build/html/_static/basic.css new file mode 100644 index 000000000..19ced1057 --- /dev/null +++ b/docs/_build/html/_static/basic.css @@ -0,0 +1,665 @@ +/* + * basic.css + * ~~~~~~~~~ + * + * Sphinx stylesheet -- basic theme. + * + * :copyright: Copyright 2007-2018 by the Sphinx team, see AUTHORS. + * :license: BSD, see LICENSE for details. + * + */ + +/* -- main layout ----------------------------------------------------------- */ + +div.clearer { + clear: both; +} + +/* -- relbar ---------------------------------------------------------------- */ + +div.related { + width: 100%; + font-size: 90%; +} + +div.related h3 { + display: none; +} + +div.related ul { + margin: 0; + padding: 0 0 0 10px; + list-style: none; +} + +div.related li { + display: inline; +} + +div.related li.right { + float: right; + margin-right: 5px; +} + +/* -- sidebar --------------------------------------------------------------- */ + +div.sphinxsidebarwrapper { + padding: 10px 5px 0 10px; +} + +div.sphinxsidebar { + float: left; + width: 230px; + margin-left: -100%; + font-size: 90%; + word-wrap: break-word; + overflow-wrap : break-word; +} + +div.sphinxsidebar ul { + list-style: none; +} + +div.sphinxsidebar ul ul, +div.sphinxsidebar ul.want-points { + margin-left: 20px; + list-style: square; +} + +div.sphinxsidebar ul ul { + margin-top: 0; + margin-bottom: 0; +} + +div.sphinxsidebar form { + margin-top: 10px; +} + +div.sphinxsidebar input { + border: 1px solid #98dbcc; + font-family: sans-serif; + font-size: 1em; +} + +div.sphinxsidebar #searchbox input[type="text"] { + float: left; + width: 80%; + padding: 0.25em; + box-sizing: border-box; +} + +div.sphinxsidebar #searchbox input[type="submit"] { + float: left; + width: 20%; + border-left: none; + padding: 0.25em; + box-sizing: border-box; +} + + +img { + border: 0; + max-width: 100%; +} + +/* -- search page ----------------------------------------------------------- */ + +ul.search { + margin: 10px 0 0 20px; + padding: 0; +} + +ul.search li { + padding: 5px 0 5px 20px; + background-image: url(file.png); + background-repeat: no-repeat; + background-position: 0 7px; +} + +ul.search li a { + font-weight: bold; +} + +ul.search li div.context { + color: #888; + margin: 2px 0 0 30px; + text-align: left; +} + +ul.keywordmatches li.goodmatch a { + font-weight: bold; +} + +/* -- index page ------------------------------------------------------------ */ + +table.contentstable { + width: 90%; + margin-left: auto; + margin-right: auto; +} + +table.contentstable p.biglink { + line-height: 150%; +} + +a.biglink { + font-size: 1.3em; +} + +span.linkdescr { + font-style: italic; + padding-top: 5px; + font-size: 90%; +} + +/* -- general index --------------------------------------------------------- */ + +table.indextable { + width: 100%; +} + +table.indextable td { + text-align: left; + vertical-align: top; +} + +table.indextable ul { + margin-top: 0; + margin-bottom: 0; + list-style-type: none; +} + +table.indextable > tbody > tr > td > ul { + padding-left: 0em; +} + +table.indextable tr.pcap { + height: 10px; +} + +table.indextable tr.cap { + margin-top: 10px; + background-color: #f2f2f2; +} + +img.toggler { + margin-right: 3px; + margin-top: 3px; + cursor: pointer; +} + +div.modindex-jumpbox { + border-top: 1px solid #ddd; + border-bottom: 1px solid #ddd; + margin: 1em 0 1em 0; + padding: 0.4em; +} + +div.genindex-jumpbox { + border-top: 1px solid #ddd; + border-bottom: 1px solid #ddd; + margin: 1em 0 1em 0; + padding: 0.4em; +} + +/* -- domain module index --------------------------------------------------- */ + +table.modindextable td { + padding: 2px; + border-collapse: collapse; +} + +/* -- general body styles --------------------------------------------------- */ + +div.body { + min-width: 450px; + max-width: 800px; +} + +div.body p, div.body dd, div.body li, div.body blockquote { + -moz-hyphens: auto; + -ms-hyphens: auto; + -webkit-hyphens: auto; + hyphens: auto; +} + +a.headerlink { + visibility: hidden; +} + +h1:hover > a.headerlink, +h2:hover > a.headerlink, +h3:hover > a.headerlink, +h4:hover > a.headerlink, +h5:hover > a.headerlink, +h6:hover > a.headerlink, +dt:hover > a.headerlink, +caption:hover > a.headerlink, +p.caption:hover > a.headerlink, +div.code-block-caption:hover > a.headerlink { + visibility: visible; +} + +div.body p.caption { + text-align: inherit; +} + +div.body td { + text-align: left; +} + +.first { + margin-top: 0 !important; +} + +p.rubric { + margin-top: 30px; + font-weight: bold; +} + +img.align-left, .figure.align-left, object.align-left { + clear: left; + float: left; + margin-right: 1em; +} + +img.align-right, .figure.align-right, object.align-right { + clear: right; + float: right; + margin-left: 1em; +} + +img.align-center, .figure.align-center, object.align-center { + display: block; + margin-left: auto; + margin-right: auto; +} + +.align-left { + text-align: left; +} + +.align-center { + text-align: center; +} + +.align-right { + text-align: right; +} + +/* -- sidebars -------------------------------------------------------------- */ + +div.sidebar { + margin: 0 0 0.5em 1em; + border: 1px solid #ddb; + padding: 7px 7px 0 7px; + background-color: #ffe; + width: 40%; + float: right; +} + +p.sidebar-title { + font-weight: bold; +} + +/* -- topics ---------------------------------------------------------------- */ + +div.topic { + border: 1px solid #ccc; + padding: 7px 7px 0 7px; + margin: 10px 0 10px 0; +} + +p.topic-title { + font-size: 1.1em; + font-weight: bold; + margin-top: 10px; +} + +/* -- admonitions ----------------------------------------------------------- */ + +div.admonition { + margin-top: 10px; + margin-bottom: 10px; + padding: 7px; +} + +div.admonition dt { + font-weight: bold; +} + +div.admonition dl { + margin-bottom: 0; +} + +p.admonition-title { + margin: 0px 10px 5px 0px; + font-weight: bold; +} + +div.body p.centered { + text-align: center; + margin-top: 25px; +} + +/* -- tables ---------------------------------------------------------------- */ + +table.docutils { + border: 0; + border-collapse: collapse; +} + +table.align-center { + margin-left: auto; + margin-right: auto; +} + +table caption span.caption-number { + font-style: italic; +} + +table caption span.caption-text { +} + +table.docutils td, table.docutils th { + padding: 1px 8px 1px 5px; + border-top: 0; + border-left: 0; + border-right: 0; + border-bottom: 1px solid #aaa; +} + +table.footnote td, table.footnote th { + border: 0 !important; +} + +th { + text-align: left; + padding-right: 5px; +} + +table.citation { + border-left: solid 1px gray; + margin-left: 1px; +} + +table.citation td { + border-bottom: none; +} + +/* -- figures --------------------------------------------------------------- */ + +div.figure { + margin: 0.5em; + padding: 0.5em; +} + +div.figure p.caption { + padding: 0.3em; +} + +div.figure p.caption span.caption-number { + font-style: italic; +} + +div.figure p.caption span.caption-text { +} + +/* -- field list styles ----------------------------------------------------- */ + +table.field-list td, table.field-list th { + border: 0 !important; +} + +.field-list ul { + margin: 0; + padding-left: 1em; +} + +.field-list p { + margin: 0; +} + +.field-name { + -moz-hyphens: manual; + -ms-hyphens: manual; + -webkit-hyphens: manual; + hyphens: manual; +} + +/* -- other body styles ----------------------------------------------------- */ + +ol.arabic { + list-style: decimal; +} + +ol.loweralpha { + list-style: lower-alpha; +} + +ol.upperalpha { + list-style: upper-alpha; +} + +ol.lowerroman { + list-style: lower-roman; +} + +ol.upperroman { + list-style: upper-roman; +} + +dl { + margin-bottom: 15px; +} + +dd p { + margin-top: 0px; +} + +dd ul, dd table { + margin-bottom: 10px; +} + +dd { + margin-top: 3px; + margin-bottom: 10px; + margin-left: 30px; +} + +dt:target, span.highlighted { + background-color: #fbe54e; +} + +rect.highlighted { + fill: #fbe54e; +} + +dl.glossary dt { + font-weight: bold; + font-size: 1.1em; +} + +.optional { + font-size: 1.3em; +} + +.sig-paren { + font-size: larger; +} + +.versionmodified { + font-style: italic; +} + +.system-message { + background-color: #fda; + padding: 5px; + border: 3px solid red; +} + +.footnote:target { + background-color: #ffa; +} + +.line-block { + display: block; + margin-top: 1em; + margin-bottom: 1em; +} + +.line-block .line-block { + margin-top: 0; + margin-bottom: 0; + margin-left: 1.5em; +} + +.guilabel, .menuselection { + font-family: sans-serif; +} + +.accelerator { + text-decoration: underline; +} + +.classifier { + font-style: oblique; +} + +abbr, acronym { + border-bottom: dotted 1px; + cursor: help; +} + +/* -- code displays --------------------------------------------------------- */ + +pre { + overflow: auto; + overflow-y: hidden; /* fixes display issues on Chrome browsers */ +} + +span.pre { + -moz-hyphens: none; + -ms-hyphens: none; + -webkit-hyphens: none; + hyphens: none; +} + +td.linenos pre { + padding: 5px 0px; + border: 0; + background-color: transparent; + color: #aaa; +} + +table.highlighttable { + margin-left: 0.5em; +} + +table.highlighttable td { + padding: 0 0.5em 0 0.5em; +} + +div.code-block-caption { + padding: 2px 5px; + font-size: small; +} + +div.code-block-caption code { + background-color: transparent; +} + +div.code-block-caption + div > div.highlight > pre { + margin-top: 0; +} + +div.code-block-caption span.caption-number { + padding: 0.1em 0.3em; + font-style: italic; +} + +div.code-block-caption span.caption-text { +} + +div.literal-block-wrapper { + padding: 1em 1em 0; +} + +div.literal-block-wrapper div.highlight { + margin: 0; +} + +code.descname { + background-color: transparent; + font-weight: bold; + font-size: 1.2em; +} + +code.descclassname { + background-color: transparent; +} + +code.xref, a code { + background-color: transparent; + font-weight: bold; +} + +h1 code, h2 code, h3 code, h4 code, h5 code, h6 code { + background-color: transparent; +} + +.viewcode-link { + float: right; +} + +.viewcode-back { + float: right; + font-family: sans-serif; +} + +div.viewcode-block:target { + margin: -1px -10px; + padding: 0 10px; +} + +/* -- math display ---------------------------------------------------------- */ + +img.math { + vertical-align: middle; +} + +div.body div.math p { + text-align: center; +} + +span.eqno { + float: right; +} + +span.eqno a.headerlink { + position: relative; + left: 0px; + z-index: 1; +} + +div.math:hover a.headerlink { + visibility: visible; +} + +/* -- printout stylesheet --------------------------------------------------- */ + +@media print { + div.document, + div.documentwrapper, + div.bodywrapper { + margin: 0 !important; + width: 100%; + } + + div.sphinxsidebar, + div.related, + div.footer, + #top-link { + display: none; + } +} \ No newline at end of file diff --git a/docs/_build/html/_static/comment-bright.png b/docs/_build/html/_static/comment-bright.png new file mode 100644 index 000000000..15e27edb1 Binary files /dev/null and b/docs/_build/html/_static/comment-bright.png differ diff --git a/docs/_build/html/_static/comment-close.png b/docs/_build/html/_static/comment-close.png new file mode 100644 index 000000000..4d91bcf57 Binary files /dev/null and b/docs/_build/html/_static/comment-close.png differ diff --git a/docs/_build/html/_static/comment.png b/docs/_build/html/_static/comment.png new file mode 100644 index 000000000..dfbc0cbd5 Binary files /dev/null and b/docs/_build/html/_static/comment.png differ diff --git a/docs/_build/html/_static/css/badge_only.css b/docs/_build/html/_static/css/badge_only.css new file mode 100644 index 000000000..323730ae2 --- /dev/null +++ b/docs/_build/html/_static/css/badge_only.css @@ -0,0 +1 @@ +.fa:before{-webkit-font-smoothing:antialiased}.clearfix{*zoom:1}.clearfix:before,.clearfix:after{display:table;content:""}.clearfix:after{clear:both}@font-face{font-family:FontAwesome;font-weight:normal;font-style:normal;src:url("../fonts/fontawesome-webfont.eot");src:url("../fonts/fontawesome-webfont.eot?#iefix") format("embedded-opentype"),url("../fonts/fontawesome-webfont.woff") format("woff"),url("../fonts/fontawesome-webfont.ttf") format("truetype"),url("../fonts/fontawesome-webfont.svg#FontAwesome") format("svg")}.fa:before{display:inline-block;font-family:FontAwesome;font-style:normal;font-weight:normal;line-height:1;text-decoration:inherit}a .fa{display:inline-block;text-decoration:inherit}li .fa{display:inline-block}li .fa-large:before,li .fa-large:before{width:1.875em}ul.fas{list-style-type:none;margin-left:2em;text-indent:-0.8em}ul.fas li .fa{width:.8em}ul.fas li .fa-large:before,ul.fas li .fa-large:before{vertical-align:baseline}.fa-book:before{content:""}.icon-book:before{content:""}.fa-caret-down:before{content:""}.icon-caret-down:before{content:""}.fa-caret-up:before{content:""}.icon-caret-up:before{content:""}.fa-caret-left:before{content:""}.icon-caret-left:before{content:""}.fa-caret-right:before{content:""}.icon-caret-right:before{content:""}.rst-versions{position:fixed;bottom:0;left:0;width:300px;color:#fcfcfc;background:#1f1d1d;font-family:"Lato","proxima-nova","Helvetica Neue",Arial,sans-serif;z-index:400}.rst-versions a{color:#2980B9;text-decoration:none}.rst-versions .rst-badge-small{display:none}.rst-versions .rst-current-version{padding:12px;background-color:#272525;display:block;text-align:right;font-size:90%;cursor:pointer;color:#27AE60;*zoom:1}.rst-versions .rst-current-version:before,.rst-versions .rst-current-version:after{display:table;content:""}.rst-versions .rst-current-version:after{clear:both}.rst-versions .rst-current-version .fa{color:#fcfcfc}.rst-versions .rst-current-version .fa-book{float:left}.rst-versions .rst-current-version .icon-book{float:left}.rst-versions .rst-current-version.rst-out-of-date{background-color:#E74C3C;color:#fff}.rst-versions .rst-current-version.rst-active-old-version{background-color:#F1C40F;color:#000}.rst-versions.shift-up{height:auto;max-height:100%}.rst-versions.shift-up .rst-other-versions{display:block}.rst-versions .rst-other-versions{font-size:90%;padding:12px;color:gray;display:none}.rst-versions .rst-other-versions hr{display:block;height:1px;border:0;margin:20px 0;padding:0;border-top:solid 1px #413d3d}.rst-versions .rst-other-versions dd{display:inline-block;margin:0}.rst-versions .rst-other-versions dd a{display:inline-block;padding:6px;color:#fcfcfc}.rst-versions.rst-badge{width:auto;bottom:20px;right:20px;left:auto;border:none;max-width:300px}.rst-versions.rst-badge .icon-book{float:none}.rst-versions.rst-badge .fa-book{float:none}.rst-versions.rst-badge.shift-up .rst-current-version{text-align:right}.rst-versions.rst-badge.shift-up .rst-current-version .fa-book{float:left}.rst-versions.rst-badge.shift-up .rst-current-version .icon-book{float:left}.rst-versions.rst-badge .rst-current-version{width:auto;height:30px;line-height:30px;padding:0 6px;display:block;text-align:center}@media screen and (max-width: 768px){.rst-versions{width:85%;display:none}.rst-versions.shift{display:block}} diff --git a/docs/_build/html/_static/css/theme.css b/docs/_build/html/_static/css/theme.css new file mode 100644 index 000000000..03a13df62 --- /dev/null +++ b/docs/_build/html/_static/css/theme.css @@ -0,0 +1,6 @@ +/* sphinx_rtd_theme version 0.4.1 | MIT license */ +/* Built 20180727 10:07 */ +*{-webkit-box-sizing:border-box;-moz-box-sizing:border-box;box-sizing:border-box}article,aside,details,figcaption,figure,footer,header,hgroup,nav,section{display:block}audio,canvas,video{display:inline-block;*display:inline;*zoom:1}audio:not([controls]){display:none}[hidden]{display:none}*{-webkit-box-sizing:border-box;-moz-box-sizing:border-box;box-sizing:border-box}html{font-size:100%;-webkit-text-size-adjust:100%;-ms-text-size-adjust:100%}body{margin:0}a:hover,a:active{outline:0}abbr[title]{border-bottom:1px dotted}b,strong{font-weight:bold}blockquote{margin:0}dfn{font-style:italic}ins{background:#ff9;color:#000;text-decoration:none}mark{background:#ff0;color:#000;font-style:italic;font-weight:bold}pre,code,.rst-content tt,.rst-content code,kbd,samp{font-family:monospace,serif;_font-family:"courier new",monospace;font-size:1em}pre{white-space:pre}q{quotes:none}q:before,q:after{content:"";content:none}small{font-size:85%}sub,sup{font-size:75%;line-height:0;position:relative;vertical-align:baseline}sup{top:-0.5em}sub{bottom:-0.25em}ul,ol,dl{margin:0;padding:0;list-style:none;list-style-image:none}li{list-style:none}dd{margin:0}img{border:0;-ms-interpolation-mode:bicubic;vertical-align:middle;max-width:100%}svg:not(:root){overflow:hidden}figure{margin:0}form{margin:0}fieldset{border:0;margin:0;padding:0}label{cursor:pointer}legend{border:0;*margin-left:-7px;padding:0;white-space:normal}button,input,select,textarea{font-size:100%;margin:0;vertical-align:baseline;*vertical-align:middle}button,input{line-height:normal}button,input[type="button"],input[type="reset"],input[type="submit"]{cursor:pointer;-webkit-appearance:button;*overflow:visible}button[disabled],input[disabled]{cursor:default}input[type="checkbox"],input[type="radio"]{box-sizing:border-box;padding:0;*width:13px;*height:13px}input[type="search"]{-webkit-appearance:textfield;-moz-box-sizing:content-box;-webkit-box-sizing:content-box;box-sizing:content-box}input[type="search"]::-webkit-search-decoration,input[type="search"]::-webkit-search-cancel-button{-webkit-appearance:none}button::-moz-focus-inner,input::-moz-focus-inner{border:0;padding:0}textarea{overflow:auto;vertical-align:top;resize:vertical}table{border-collapse:collapse;border-spacing:0}td{vertical-align:top}.chromeframe{margin:.2em 0;background:#ccc;color:#000;padding:.2em 0}.ir{display:block;border:0;text-indent:-999em;overflow:hidden;background-color:transparent;background-repeat:no-repeat;text-align:left;direction:ltr;*line-height:0}.ir br{display:none}.hidden{display:none !important;visibility:hidden}.visuallyhidden{border:0;clip:rect(0 0 0 0);height:1px;margin:-1px;overflow:hidden;padding:0;position:absolute;width:1px}.visuallyhidden.focusable:active,.visuallyhidden.focusable:focus{clip:auto;height:auto;margin:0;overflow:visible;position:static;width:auto}.invisible{visibility:hidden}.relative{position:relative}big,small{font-size:100%}@media print{html,body,section{background:none !important}*{box-shadow:none !important;text-shadow:none !important;filter:none !important;-ms-filter:none !important}a,a:visited{text-decoration:underline}.ir a:after,a[href^="javascript:"]:after,a[href^="#"]:after{content:""}pre,blockquote{page-break-inside:avoid}thead{display:table-header-group}tr,img{page-break-inside:avoid}img{max-width:100% !important}@page{margin:.5cm}p,h2,.rst-content .toctree-wrapper p.caption,h3{orphans:3;widows:3}h2,.rst-content .toctree-wrapper p.caption,h3{page-break-after:avoid}}.fa:before,.wy-menu-vertical li span.toctree-expand:before,.wy-menu-vertical li.on a span.toctree-expand:before,.wy-menu-vertical li.current>a span.toctree-expand:before,.rst-content .admonition-title:before,.rst-content h1 .headerlink:before,.rst-content h2 .headerlink:before,.rst-content h3 .headerlink:before,.rst-content h4 .headerlink:before,.rst-content h5 .headerlink:before,.rst-content h6 .headerlink:before,.rst-content dl dt .headerlink:before,.rst-content p.caption .headerlink:before,.rst-content table>caption .headerlink:before,.rst-content tt.download span:first-child:before,.rst-content code.download span:first-child:before,.icon:before,.wy-dropdown .caret:before,.wy-inline-validate.wy-inline-validate-success .wy-input-context:before,.wy-inline-validate.wy-inline-validate-danger .wy-input-context:before,.wy-inline-validate.wy-inline-validate-warning .wy-input-context:before,.wy-inline-validate.wy-inline-validate-info .wy-input-context:before,.wy-alert,.rst-content .note,.rst-content .attention,.rst-content .caution,.rst-content .danger,.rst-content .error,.rst-content .hint,.rst-content .important,.rst-content .tip,.rst-content .warning,.rst-content .seealso,.rst-content .admonition-todo,.rst-content .admonition,.btn,input[type="text"],input[type="password"],input[type="email"],input[type="url"],input[type="date"],input[type="month"],input[type="time"],input[type="datetime"],input[type="datetime-local"],input[type="week"],input[type="number"],input[type="search"],input[type="tel"],input[type="color"],select,textarea,.wy-menu-vertical li.on a,.wy-menu-vertical li.current>a,.wy-side-nav-search>a,.wy-side-nav-search .wy-dropdown>a,.wy-nav-top a{-webkit-font-smoothing:antialiased}.clearfix{*zoom:1}.clearfix:before,.clearfix:after{display:table;content:""}.clearfix:after{clear:both}/*! + * Font Awesome 4.7.0 by @davegandy - http://fontawesome.io - @fontawesome + * License - http://fontawesome.io/license (Font: SIL OFL 1.1, CSS: MIT License) + */@font-face{font-family:'FontAwesome';src:url("../fonts/fontawesome-webfont.eot?v=4.7.0");src:url("../fonts/fontawesome-webfont.eot?#iefix&v=4.7.0") format("embedded-opentype"),url("../fonts/fontawesome-webfont.woff2?v=4.7.0") format("woff2"),url("../fonts/fontawesome-webfont.woff?v=4.7.0") format("woff"),url("../fonts/fontawesome-webfont.ttf?v=4.7.0") format("truetype"),url("../fonts/fontawesome-webfont.svg?v=4.7.0#fontawesomeregular") format("svg");font-weight:normal;font-style:normal}.fa,.wy-menu-vertical li span.toctree-expand,.wy-menu-vertical li.on a span.toctree-expand,.wy-menu-vertical li.current>a span.toctree-expand,.rst-content .admonition-title,.rst-content h1 .headerlink,.rst-content h2 .headerlink,.rst-content h3 .headerlink,.rst-content h4 .headerlink,.rst-content h5 .headerlink,.rst-content h6 .headerlink,.rst-content dl dt .headerlink,.rst-content p.caption .headerlink,.rst-content table>caption .headerlink,.rst-content tt.download span:first-child,.rst-content code.download span:first-child,.icon{display:inline-block;font:normal normal normal 14px/1 FontAwesome;font-size:inherit;text-rendering:auto;-webkit-font-smoothing:antialiased;-moz-osx-font-smoothing:grayscale}.fa-lg{font-size:1.3333333333em;line-height:.75em;vertical-align:-15%}.fa-2x{font-size:2em}.fa-3x{font-size:3em}.fa-4x{font-size:4em}.fa-5x{font-size:5em}.fa-fw{width:1.2857142857em;text-align:center}.fa-ul{padding-left:0;margin-left:2.1428571429em;list-style-type:none}.fa-ul>li{position:relative}.fa-li{position:absolute;left:-2.1428571429em;width:2.1428571429em;top:.1428571429em;text-align:center}.fa-li.fa-lg{left:-1.8571428571em}.fa-border{padding:.2em .25em .15em;border:solid 0.08em #eee;border-radius:.1em}.fa-pull-left{float:left}.fa-pull-right{float:right}.fa.fa-pull-left,.wy-menu-vertical li span.fa-pull-left.toctree-expand,.wy-menu-vertical li.on a span.fa-pull-left.toctree-expand,.wy-menu-vertical li.current>a span.fa-pull-left.toctree-expand,.rst-content .fa-pull-left.admonition-title,.rst-content h1 .fa-pull-left.headerlink,.rst-content h2 .fa-pull-left.headerlink,.rst-content h3 .fa-pull-left.headerlink,.rst-content h4 .fa-pull-left.headerlink,.rst-content h5 .fa-pull-left.headerlink,.rst-content h6 .fa-pull-left.headerlink,.rst-content dl dt .fa-pull-left.headerlink,.rst-content p.caption .fa-pull-left.headerlink,.rst-content table>caption .fa-pull-left.headerlink,.rst-content tt.download span.fa-pull-left:first-child,.rst-content code.download span.fa-pull-left:first-child,.fa-pull-left.icon{margin-right:.3em}.fa.fa-pull-right,.wy-menu-vertical li span.fa-pull-right.toctree-expand,.wy-menu-vertical li.on a span.fa-pull-right.toctree-expand,.wy-menu-vertical li.current>a span.fa-pull-right.toctree-expand,.rst-content .fa-pull-right.admonition-title,.rst-content h1 .fa-pull-right.headerlink,.rst-content h2 .fa-pull-right.headerlink,.rst-content h3 .fa-pull-right.headerlink,.rst-content h4 .fa-pull-right.headerlink,.rst-content h5 .fa-pull-right.headerlink,.rst-content h6 .fa-pull-right.headerlink,.rst-content dl dt .fa-pull-right.headerlink,.rst-content p.caption .fa-pull-right.headerlink,.rst-content table>caption .fa-pull-right.headerlink,.rst-content tt.download span.fa-pull-right:first-child,.rst-content code.download span.fa-pull-right:first-child,.fa-pull-right.icon{margin-left:.3em}.pull-right{float:right}.pull-left{float:left}.fa.pull-left,.wy-menu-vertical li span.pull-left.toctree-expand,.wy-menu-vertical li.on a span.pull-left.toctree-expand,.wy-menu-vertical li.current>a span.pull-left.toctree-expand,.rst-content .pull-left.admonition-title,.rst-content h1 .pull-left.headerlink,.rst-content h2 .pull-left.headerlink,.rst-content h3 .pull-left.headerlink,.rst-content h4 .pull-left.headerlink,.rst-content h5 .pull-left.headerlink,.rst-content h6 .pull-left.headerlink,.rst-content dl dt .pull-left.headerlink,.rst-content p.caption .pull-left.headerlink,.rst-content table>caption .pull-left.headerlink,.rst-content tt.download span.pull-left:first-child,.rst-content code.download span.pull-left:first-child,.pull-left.icon{margin-right:.3em}.fa.pull-right,.wy-menu-vertical li span.pull-right.toctree-expand,.wy-menu-vertical li.on a span.pull-right.toctree-expand,.wy-menu-vertical li.current>a span.pull-right.toctree-expand,.rst-content .pull-right.admonition-title,.rst-content h1 .pull-right.headerlink,.rst-content h2 .pull-right.headerlink,.rst-content h3 .pull-right.headerlink,.rst-content h4 .pull-right.headerlink,.rst-content h5 .pull-right.headerlink,.rst-content h6 .pull-right.headerlink,.rst-content dl dt .pull-right.headerlink,.rst-content p.caption .pull-right.headerlink,.rst-content table>caption .pull-right.headerlink,.rst-content tt.download span.pull-right:first-child,.rst-content code.download span.pull-right:first-child,.pull-right.icon{margin-left:.3em}.fa-spin{-webkit-animation:fa-spin 2s infinite linear;animation:fa-spin 2s infinite linear}.fa-pulse{-webkit-animation:fa-spin 1s infinite steps(8);animation:fa-spin 1s infinite steps(8)}@-webkit-keyframes fa-spin{0%{-webkit-transform:rotate(0deg);transform:rotate(0deg)}100%{-webkit-transform:rotate(359deg);transform:rotate(359deg)}}@keyframes fa-spin{0%{-webkit-transform:rotate(0deg);transform:rotate(0deg)}100%{-webkit-transform:rotate(359deg);transform:rotate(359deg)}}.fa-rotate-90{-ms-filter:"progid:DXImageTransform.Microsoft.BasicImage(rotation=1)";-webkit-transform:rotate(90deg);-ms-transform:rotate(90deg);transform:rotate(90deg)}.fa-rotate-180{-ms-filter:"progid:DXImageTransform.Microsoft.BasicImage(rotation=2)";-webkit-transform:rotate(180deg);-ms-transform:rotate(180deg);transform:rotate(180deg)}.fa-rotate-270{-ms-filter:"progid:DXImageTransform.Microsoft.BasicImage(rotation=3)";-webkit-transform:rotate(270deg);-ms-transform:rotate(270deg);transform:rotate(270deg)}.fa-flip-horizontal{-ms-filter:"progid:DXImageTransform.Microsoft.BasicImage(rotation=0, mirror=1)";-webkit-transform:scale(-1, 1);-ms-transform:scale(-1, 1);transform:scale(-1, 1)}.fa-flip-vertical{-ms-filter:"progid:DXImageTransform.Microsoft.BasicImage(rotation=2, mirror=1)";-webkit-transform:scale(1, -1);-ms-transform:scale(1, -1);transform:scale(1, -1)}:root .fa-rotate-90,:root .fa-rotate-180,:root .fa-rotate-270,:root .fa-flip-horizontal,:root .fa-flip-vertical{filter:none}.fa-stack{position:relative;display:inline-block;width:2em;height:2em;line-height:2em;vertical-align:middle}.fa-stack-1x,.fa-stack-2x{position:absolute;left:0;width:100%;text-align:center}.fa-stack-1x{line-height:inherit}.fa-stack-2x{font-size:2em}.fa-inverse{color:#fff}.fa-glass:before{content:""}.fa-music:before{content:""}.fa-search:before,.icon-search:before{content:""}.fa-envelope-o:before{content:""}.fa-heart:before{content:""}.fa-star:before{content:""}.fa-star-o:before{content:""}.fa-user:before{content:""}.fa-film:before{content:""}.fa-th-large:before{content:""}.fa-th:before{content:""}.fa-th-list:before{content:""}.fa-check:before{content:""}.fa-remove:before,.fa-close:before,.fa-times:before{content:""}.fa-search-plus:before{content:""}.fa-search-minus:before{content:""}.fa-power-off:before{content:""}.fa-signal:before{content:""}.fa-gear:before,.fa-cog:before{content:""}.fa-trash-o:before{content:""}.fa-home:before,.icon-home:before{content:""}.fa-file-o:before{content:""}.fa-clock-o:before{content:""}.fa-road:before{content:""}.fa-download:before,.rst-content tt.download span:first-child:before,.rst-content code.download span:first-child:before{content:""}.fa-arrow-circle-o-down:before{content:""}.fa-arrow-circle-o-up:before{content:""}.fa-inbox:before{content:""}.fa-play-circle-o:before{content:""}.fa-rotate-right:before,.fa-repeat:before{content:""}.fa-refresh:before{content:""}.fa-list-alt:before{content:""}.fa-lock:before{content:""}.fa-flag:before{content:""}.fa-headphones:before{content:""}.fa-volume-off:before{content:""}.fa-volume-down:before{content:""}.fa-volume-up:before{content:""}.fa-qrcode:before{content:""}.fa-barcode:before{content:""}.fa-tag:before{content:""}.fa-tags:before{content:""}.fa-book:before,.icon-book:before{content:""}.fa-bookmark:before{content:""}.fa-print:before{content:""}.fa-camera:before{content:""}.fa-font:before{content:""}.fa-bold:before{content:""}.fa-italic:before{content:""}.fa-text-height:before{content:""}.fa-text-width:before{content:""}.fa-align-left:before{content:""}.fa-align-center:before{content:""}.fa-align-right:before{content:""}.fa-align-justify:before{content:""}.fa-list:before{content:""}.fa-dedent:before,.fa-outdent:before{content:""}.fa-indent:before{content:""}.fa-video-camera:before{content:""}.fa-photo:before,.fa-image:before,.fa-picture-o:before{content:""}.fa-pencil:before{content:""}.fa-map-marker:before{content:""}.fa-adjust:before{content:""}.fa-tint:before{content:""}.fa-edit:before,.fa-pencil-square-o:before{content:""}.fa-share-square-o:before{content:""}.fa-check-square-o:before{content:""}.fa-arrows:before{content:""}.fa-step-backward:before{content:""}.fa-fast-backward:before{content:""}.fa-backward:before{content:""}.fa-play:before{content:""}.fa-pause:before{content:""}.fa-stop:before{content:""}.fa-forward:before{content:""}.fa-fast-forward:before{content:""}.fa-step-forward:before{content:""}.fa-eject:before{content:""}.fa-chevron-left:before{content:""}.fa-chevron-right:before{content:""}.fa-plus-circle:before{content:""}.fa-minus-circle:before{content:""}.fa-times-circle:before,.wy-inline-validate.wy-inline-validate-danger .wy-input-context:before{content:""}.fa-check-circle:before,.wy-inline-validate.wy-inline-validate-success .wy-input-context:before{content:""}.fa-question-circle:before{content:""}.fa-info-circle:before{content:""}.fa-crosshairs:before{content:""}.fa-times-circle-o:before{content:""}.fa-check-circle-o:before{content:""}.fa-ban:before{content:""}.fa-arrow-left:before{content:""}.fa-arrow-right:before{content:""}.fa-arrow-up:before{content:""}.fa-arrow-down:before{content:""}.fa-mail-forward:before,.fa-share:before{content:""}.fa-expand:before{content:""}.fa-compress:before{content:""}.fa-plus:before{content:""}.fa-minus:before{content:""}.fa-asterisk:before{content:""}.fa-exclamation-circle:before,.wy-inline-validate.wy-inline-validate-warning .wy-input-context:before,.wy-inline-validate.wy-inline-validate-info .wy-input-context:before,.rst-content .admonition-title:before{content:""}.fa-gift:before{content:""}.fa-leaf:before{content:""}.fa-fire:before,.icon-fire:before{content:""}.fa-eye:before{content:""}.fa-eye-slash:before{content:""}.fa-warning:before,.fa-exclamation-triangle:before{content:""}.fa-plane:before{content:""}.fa-calendar:before{content:""}.fa-random:before{content:""}.fa-comment:before{content:""}.fa-magnet:before{content:""}.fa-chevron-up:before{content:""}.fa-chevron-down:before{content:""}.fa-retweet:before{content:""}.fa-shopping-cart:before{content:""}.fa-folder:before{content:""}.fa-folder-open:before{content:""}.fa-arrows-v:before{content:""}.fa-arrows-h:before{content:""}.fa-bar-chart-o:before,.fa-bar-chart:before{content:""}.fa-twitter-square:before{content:""}.fa-facebook-square:before{content:""}.fa-camera-retro:before{content:""}.fa-key:before{content:""}.fa-gears:before,.fa-cogs:before{content:""}.fa-comments:before{content:""}.fa-thumbs-o-up:before{content:""}.fa-thumbs-o-down:before{content:""}.fa-star-half:before{content:""}.fa-heart-o:before{content:""}.fa-sign-out:before{content:""}.fa-linkedin-square:before{content:""}.fa-thumb-tack:before{content:""}.fa-external-link:before{content:""}.fa-sign-in:before{content:""}.fa-trophy:before{content:""}.fa-github-square:before{content:""}.fa-upload:before{content:""}.fa-lemon-o:before{content:""}.fa-phone:before{content:""}.fa-square-o:before{content:""}.fa-bookmark-o:before{content:""}.fa-phone-square:before{content:""}.fa-twitter:before{content:""}.fa-facebook-f:before,.fa-facebook:before{content:""}.fa-github:before,.icon-github:before{content:""}.fa-unlock:before{content:""}.fa-credit-card:before{content:""}.fa-feed:before,.fa-rss:before{content:""}.fa-hdd-o:before{content:""}.fa-bullhorn:before{content:""}.fa-bell:before{content:""}.fa-certificate:before{content:""}.fa-hand-o-right:before{content:""}.fa-hand-o-left:before{content:""}.fa-hand-o-up:before{content:""}.fa-hand-o-down:before{content:""}.fa-arrow-circle-left:before,.icon-circle-arrow-left:before{content:""}.fa-arrow-circle-right:before,.icon-circle-arrow-right:before{content:""}.fa-arrow-circle-up:before{content:""}.fa-arrow-circle-down:before{content:""}.fa-globe:before{content:""}.fa-wrench:before{content:""}.fa-tasks:before{content:""}.fa-filter:before{content:""}.fa-briefcase:before{content:""}.fa-arrows-alt:before{content:""}.fa-group:before,.fa-users:before{content:""}.fa-chain:before,.fa-link:before,.icon-link:before{content:""}.fa-cloud:before{content:""}.fa-flask:before{content:""}.fa-cut:before,.fa-scissors:before{content:""}.fa-copy:before,.fa-files-o:before{content:""}.fa-paperclip:before{content:""}.fa-save:before,.fa-floppy-o:before{content:""}.fa-square:before{content:""}.fa-navicon:before,.fa-reorder:before,.fa-bars:before{content:""}.fa-list-ul:before{content:""}.fa-list-ol:before{content:""}.fa-strikethrough:before{content:""}.fa-underline:before{content:""}.fa-table:before{content:""}.fa-magic:before{content:""}.fa-truck:before{content:""}.fa-pinterest:before{content:""}.fa-pinterest-square:before{content:""}.fa-google-plus-square:before{content:""}.fa-google-plus:before{content:""}.fa-money:before{content:""}.fa-caret-down:before,.wy-dropdown .caret:before,.icon-caret-down:before{content:""}.fa-caret-up:before{content:""}.fa-caret-left:before{content:""}.fa-caret-right:before{content:""}.fa-columns:before{content:""}.fa-unsorted:before,.fa-sort:before{content:""}.fa-sort-down:before,.fa-sort-desc:before{content:""}.fa-sort-up:before,.fa-sort-asc:before{content:""}.fa-envelope:before{content:""}.fa-linkedin:before{content:""}.fa-rotate-left:before,.fa-undo:before{content:""}.fa-legal:before,.fa-gavel:before{content:""}.fa-dashboard:before,.fa-tachometer:before{content:""}.fa-comment-o:before{content:""}.fa-comments-o:before{content:""}.fa-flash:before,.fa-bolt:before{content:""}.fa-sitemap:before{content:""}.fa-umbrella:before{content:""}.fa-paste:before,.fa-clipboard:before{content:""}.fa-lightbulb-o:before{content:""}.fa-exchange:before{content:""}.fa-cloud-download:before{content:""}.fa-cloud-upload:before{content:""}.fa-user-md:before{content:""}.fa-stethoscope:before{content:""}.fa-suitcase:before{content:""}.fa-bell-o:before{content:""}.fa-coffee:before{content:""}.fa-cutlery:before{content:""}.fa-file-text-o:before{content:""}.fa-building-o:before{content:""}.fa-hospital-o:before{content:""}.fa-ambulance:before{content:""}.fa-medkit:before{content:""}.fa-fighter-jet:before{content:""}.fa-beer:before{content:""}.fa-h-square:before{content:""}.fa-plus-square:before{content:""}.fa-angle-double-left:before{content:""}.fa-angle-double-right:before{content:""}.fa-angle-double-up:before{content:""}.fa-angle-double-down:before{content:""}.fa-angle-left:before{content:""}.fa-angle-right:before{content:""}.fa-angle-up:before{content:""}.fa-angle-down:before{content:""}.fa-desktop:before{content:""}.fa-laptop:before{content:""}.fa-tablet:before{content:""}.fa-mobile-phone:before,.fa-mobile:before{content:""}.fa-circle-o:before{content:""}.fa-quote-left:before{content:""}.fa-quote-right:before{content:""}.fa-spinner:before{content:""}.fa-circle:before{content:""}.fa-mail-reply:before,.fa-reply:before{content:""}.fa-github-alt:before{content:""}.fa-folder-o:before{content:""}.fa-folder-open-o:before{content:""}.fa-smile-o:before{content:""}.fa-frown-o:before{content:""}.fa-meh-o:before{content:""}.fa-gamepad:before{content:""}.fa-keyboard-o:before{content:""}.fa-flag-o:before{content:""}.fa-flag-checkered:before{content:""}.fa-terminal:before{content:""}.fa-code:before{content:""}.fa-mail-reply-all:before,.fa-reply-all:before{content:""}.fa-star-half-empty:before,.fa-star-half-full:before,.fa-star-half-o:before{content:""}.fa-location-arrow:before{content:""}.fa-crop:before{content:""}.fa-code-fork:before{content:""}.fa-unlink:before,.fa-chain-broken:before{content:""}.fa-question:before{content:""}.fa-info:before{content:""}.fa-exclamation:before{content:""}.fa-superscript:before{content:""}.fa-subscript:before{content:""}.fa-eraser:before{content:""}.fa-puzzle-piece:before{content:""}.fa-microphone:before{content:""}.fa-microphone-slash:before{content:""}.fa-shield:before{content:""}.fa-calendar-o:before{content:""}.fa-fire-extinguisher:before{content:""}.fa-rocket:before{content:""}.fa-maxcdn:before{content:""}.fa-chevron-circle-left:before{content:""}.fa-chevron-circle-right:before{content:""}.fa-chevron-circle-up:before{content:""}.fa-chevron-circle-down:before{content:""}.fa-html5:before{content:""}.fa-css3:before{content:""}.fa-anchor:before{content:""}.fa-unlock-alt:before{content:""}.fa-bullseye:before{content:""}.fa-ellipsis-h:before{content:""}.fa-ellipsis-v:before{content:""}.fa-rss-square:before{content:""}.fa-play-circle:before{content:""}.fa-ticket:before{content:""}.fa-minus-square:before{content:""}.fa-minus-square-o:before,.wy-menu-vertical li.on a span.toctree-expand:before,.wy-menu-vertical li.current>a span.toctree-expand:before{content:""}.fa-level-up:before{content:""}.fa-level-down:before{content:""}.fa-check-square:before{content:""}.fa-pencil-square:before{content:""}.fa-external-link-square:before{content:""}.fa-share-square:before{content:""}.fa-compass:before{content:""}.fa-toggle-down:before,.fa-caret-square-o-down:before{content:""}.fa-toggle-up:before,.fa-caret-square-o-up:before{content:""}.fa-toggle-right:before,.fa-caret-square-o-right:before{content:""}.fa-euro:before,.fa-eur:before{content:""}.fa-gbp:before{content:""}.fa-dollar:before,.fa-usd:before{content:""}.fa-rupee:before,.fa-inr:before{content:""}.fa-cny:before,.fa-rmb:before,.fa-yen:before,.fa-jpy:before{content:""}.fa-ruble:before,.fa-rouble:before,.fa-rub:before{content:""}.fa-won:before,.fa-krw:before{content:""}.fa-bitcoin:before,.fa-btc:before{content:""}.fa-file:before{content:""}.fa-file-text:before{content:""}.fa-sort-alpha-asc:before{content:""}.fa-sort-alpha-desc:before{content:""}.fa-sort-amount-asc:before{content:""}.fa-sort-amount-desc:before{content:""}.fa-sort-numeric-asc:before{content:""}.fa-sort-numeric-desc:before{content:""}.fa-thumbs-up:before{content:""}.fa-thumbs-down:before{content:""}.fa-youtube-square:before{content:""}.fa-youtube:before{content:""}.fa-xing:before{content:""}.fa-xing-square:before{content:""}.fa-youtube-play:before{content:""}.fa-dropbox:before{content:""}.fa-stack-overflow:before{content:""}.fa-instagram:before{content:""}.fa-flickr:before{content:""}.fa-adn:before{content:""}.fa-bitbucket:before,.icon-bitbucket:before{content:""}.fa-bitbucket-square:before{content:""}.fa-tumblr:before{content:""}.fa-tumblr-square:before{content:""}.fa-long-arrow-down:before{content:""}.fa-long-arrow-up:before{content:""}.fa-long-arrow-left:before{content:""}.fa-long-arrow-right:before{content:""}.fa-apple:before{content:""}.fa-windows:before{content:""}.fa-android:before{content:""}.fa-linux:before{content:""}.fa-dribbble:before{content:""}.fa-skype:before{content:""}.fa-foursquare:before{content:""}.fa-trello:before{content:""}.fa-female:before{content:""}.fa-male:before{content:""}.fa-gittip:before,.fa-gratipay:before{content:""}.fa-sun-o:before{content:""}.fa-moon-o:before{content:""}.fa-archive:before{content:""}.fa-bug:before{content:""}.fa-vk:before{content:""}.fa-weibo:before{content:""}.fa-renren:before{content:""}.fa-pagelines:before{content:""}.fa-stack-exchange:before{content:""}.fa-arrow-circle-o-right:before{content:""}.fa-arrow-circle-o-left:before{content:""}.fa-toggle-left:before,.fa-caret-square-o-left:before{content:""}.fa-dot-circle-o:before{content:""}.fa-wheelchair:before{content:""}.fa-vimeo-square:before{content:""}.fa-turkish-lira:before,.fa-try:before{content:""}.fa-plus-square-o:before,.wy-menu-vertical li span.toctree-expand:before{content:""}.fa-space-shuttle:before{content:""}.fa-slack:before{content:""}.fa-envelope-square:before{content:""}.fa-wordpress:before{content:""}.fa-openid:before{content:""}.fa-institution:before,.fa-bank:before,.fa-university:before{content:""}.fa-mortar-board:before,.fa-graduation-cap:before{content:""}.fa-yahoo:before{content:""}.fa-google:before{content:""}.fa-reddit:before{content:""}.fa-reddit-square:before{content:""}.fa-stumbleupon-circle:before{content:""}.fa-stumbleupon:before{content:""}.fa-delicious:before{content:""}.fa-digg:before{content:""}.fa-pied-piper-pp:before{content:""}.fa-pied-piper-alt:before{content:""}.fa-drupal:before{content:""}.fa-joomla:before{content:""}.fa-language:before{content:""}.fa-fax:before{content:""}.fa-building:before{content:""}.fa-child:before{content:""}.fa-paw:before{content:""}.fa-spoon:before{content:""}.fa-cube:before{content:""}.fa-cubes:before{content:""}.fa-behance:before{content:""}.fa-behance-square:before{content:""}.fa-steam:before{content:""}.fa-steam-square:before{content:""}.fa-recycle:before{content:""}.fa-automobile:before,.fa-car:before{content:""}.fa-cab:before,.fa-taxi:before{content:""}.fa-tree:before{content:""}.fa-spotify:before{content:""}.fa-deviantart:before{content:""}.fa-soundcloud:before{content:""}.fa-database:before{content:""}.fa-file-pdf-o:before{content:""}.fa-file-word-o:before{content:""}.fa-file-excel-o:before{content:""}.fa-file-powerpoint-o:before{content:""}.fa-file-photo-o:before,.fa-file-picture-o:before,.fa-file-image-o:before{content:""}.fa-file-zip-o:before,.fa-file-archive-o:before{content:""}.fa-file-sound-o:before,.fa-file-audio-o:before{content:""}.fa-file-movie-o:before,.fa-file-video-o:before{content:""}.fa-file-code-o:before{content:""}.fa-vine:before{content:""}.fa-codepen:before{content:""}.fa-jsfiddle:before{content:""}.fa-life-bouy:before,.fa-life-buoy:before,.fa-life-saver:before,.fa-support:before,.fa-life-ring:before{content:""}.fa-circle-o-notch:before{content:""}.fa-ra:before,.fa-resistance:before,.fa-rebel:before{content:""}.fa-ge:before,.fa-empire:before{content:""}.fa-git-square:before{content:""}.fa-git:before{content:""}.fa-y-combinator-square:before,.fa-yc-square:before,.fa-hacker-news:before{content:""}.fa-tencent-weibo:before{content:""}.fa-qq:before{content:""}.fa-wechat:before,.fa-weixin:before{content:""}.fa-send:before,.fa-paper-plane:before{content:""}.fa-send-o:before,.fa-paper-plane-o:before{content:""}.fa-history:before{content:""}.fa-circle-thin:before{content:""}.fa-header:before{content:""}.fa-paragraph:before{content:""}.fa-sliders:before{content:""}.fa-share-alt:before{content:""}.fa-share-alt-square:before{content:""}.fa-bomb:before{content:""}.fa-soccer-ball-o:before,.fa-futbol-o:before{content:""}.fa-tty:before{content:""}.fa-binoculars:before{content:""}.fa-plug:before{content:""}.fa-slideshare:before{content:""}.fa-twitch:before{content:""}.fa-yelp:before{content:""}.fa-newspaper-o:before{content:""}.fa-wifi:before{content:""}.fa-calculator:before{content:""}.fa-paypal:before{content:""}.fa-google-wallet:before{content:""}.fa-cc-visa:before{content:""}.fa-cc-mastercard:before{content:""}.fa-cc-discover:before{content:""}.fa-cc-amex:before{content:""}.fa-cc-paypal:before{content:""}.fa-cc-stripe:before{content:""}.fa-bell-slash:before{content:""}.fa-bell-slash-o:before{content:""}.fa-trash:before{content:""}.fa-copyright:before{content:""}.fa-at:before{content:""}.fa-eyedropper:before{content:""}.fa-paint-brush:before{content:""}.fa-birthday-cake:before{content:""}.fa-area-chart:before{content:""}.fa-pie-chart:before{content:""}.fa-line-chart:before{content:""}.fa-lastfm:before{content:""}.fa-lastfm-square:before{content:""}.fa-toggle-off:before{content:""}.fa-toggle-on:before{content:""}.fa-bicycle:before{content:""}.fa-bus:before{content:""}.fa-ioxhost:before{content:""}.fa-angellist:before{content:""}.fa-cc:before{content:""}.fa-shekel:before,.fa-sheqel:before,.fa-ils:before{content:""}.fa-meanpath:before{content:""}.fa-buysellads:before{content:""}.fa-connectdevelop:before{content:""}.fa-dashcube:before{content:""}.fa-forumbee:before{content:""}.fa-leanpub:before{content:""}.fa-sellsy:before{content:""}.fa-shirtsinbulk:before{content:""}.fa-simplybuilt:before{content:""}.fa-skyatlas:before{content:""}.fa-cart-plus:before{content:""}.fa-cart-arrow-down:before{content:""}.fa-diamond:before{content:""}.fa-ship:before{content:""}.fa-user-secret:before{content:""}.fa-motorcycle:before{content:""}.fa-street-view:before{content:""}.fa-heartbeat:before{content:""}.fa-venus:before{content:""}.fa-mars:before{content:""}.fa-mercury:before{content:""}.fa-intersex:before,.fa-transgender:before{content:""}.fa-transgender-alt:before{content:""}.fa-venus-double:before{content:""}.fa-mars-double:before{content:""}.fa-venus-mars:before{content:""}.fa-mars-stroke:before{content:""}.fa-mars-stroke-v:before{content:""}.fa-mars-stroke-h:before{content:""}.fa-neuter:before{content:""}.fa-genderless:before{content:""}.fa-facebook-official:before{content:""}.fa-pinterest-p:before{content:""}.fa-whatsapp:before{content:""}.fa-server:before{content:""}.fa-user-plus:before{content:""}.fa-user-times:before{content:""}.fa-hotel:before,.fa-bed:before{content:""}.fa-viacoin:before{content:""}.fa-train:before{content:""}.fa-subway:before{content:""}.fa-medium:before{content:""}.fa-yc:before,.fa-y-combinator:before{content:""}.fa-optin-monster:before{content:""}.fa-opencart:before{content:""}.fa-expeditedssl:before{content:""}.fa-battery-4:before,.fa-battery:before,.fa-battery-full:before{content:""}.fa-battery-3:before,.fa-battery-three-quarters:before{content:""}.fa-battery-2:before,.fa-battery-half:before{content:""}.fa-battery-1:before,.fa-battery-quarter:before{content:""}.fa-battery-0:before,.fa-battery-empty:before{content:""}.fa-mouse-pointer:before{content:""}.fa-i-cursor:before{content:""}.fa-object-group:before{content:""}.fa-object-ungroup:before{content:""}.fa-sticky-note:before{content:""}.fa-sticky-note-o:before{content:""}.fa-cc-jcb:before{content:""}.fa-cc-diners-club:before{content:""}.fa-clone:before{content:""}.fa-balance-scale:before{content:""}.fa-hourglass-o:before{content:""}.fa-hourglass-1:before,.fa-hourglass-start:before{content:""}.fa-hourglass-2:before,.fa-hourglass-half:before{content:""}.fa-hourglass-3:before,.fa-hourglass-end:before{content:""}.fa-hourglass:before{content:""}.fa-hand-grab-o:before,.fa-hand-rock-o:before{content:""}.fa-hand-stop-o:before,.fa-hand-paper-o:before{content:""}.fa-hand-scissors-o:before{content:""}.fa-hand-lizard-o:before{content:""}.fa-hand-spock-o:before{content:""}.fa-hand-pointer-o:before{content:""}.fa-hand-peace-o:before{content:""}.fa-trademark:before{content:""}.fa-registered:before{content:""}.fa-creative-commons:before{content:""}.fa-gg:before{content:""}.fa-gg-circle:before{content:""}.fa-tripadvisor:before{content:""}.fa-odnoklassniki:before{content:""}.fa-odnoklassniki-square:before{content:""}.fa-get-pocket:before{content:""}.fa-wikipedia-w:before{content:""}.fa-safari:before{content:""}.fa-chrome:before{content:""}.fa-firefox:before{content:""}.fa-opera:before{content:""}.fa-internet-explorer:before{content:""}.fa-tv:before,.fa-television:before{content:""}.fa-contao:before{content:""}.fa-500px:before{content:""}.fa-amazon:before{content:""}.fa-calendar-plus-o:before{content:""}.fa-calendar-minus-o:before{content:""}.fa-calendar-times-o:before{content:""}.fa-calendar-check-o:before{content:""}.fa-industry:before{content:""}.fa-map-pin:before{content:""}.fa-map-signs:before{content:""}.fa-map-o:before{content:""}.fa-map:before{content:""}.fa-commenting:before{content:""}.fa-commenting-o:before{content:""}.fa-houzz:before{content:""}.fa-vimeo:before{content:""}.fa-black-tie:before{content:""}.fa-fonticons:before{content:""}.fa-reddit-alien:before{content:""}.fa-edge:before{content:""}.fa-credit-card-alt:before{content:""}.fa-codiepie:before{content:""}.fa-modx:before{content:""}.fa-fort-awesome:before{content:""}.fa-usb:before{content:""}.fa-product-hunt:before{content:""}.fa-mixcloud:before{content:""}.fa-scribd:before{content:""}.fa-pause-circle:before{content:""}.fa-pause-circle-o:before{content:""}.fa-stop-circle:before{content:""}.fa-stop-circle-o:before{content:""}.fa-shopping-bag:before{content:""}.fa-shopping-basket:before{content:""}.fa-hashtag:before{content:""}.fa-bluetooth:before{content:""}.fa-bluetooth-b:before{content:""}.fa-percent:before{content:""}.fa-gitlab:before,.icon-gitlab:before{content:""}.fa-wpbeginner:before{content:""}.fa-wpforms:before{content:""}.fa-envira:before{content:""}.fa-universal-access:before{content:""}.fa-wheelchair-alt:before{content:""}.fa-question-circle-o:before{content:""}.fa-blind:before{content:""}.fa-audio-description:before{content:""}.fa-volume-control-phone:before{content:""}.fa-braille:before{content:""}.fa-assistive-listening-systems:before{content:""}.fa-asl-interpreting:before,.fa-american-sign-language-interpreting:before{content:""}.fa-deafness:before,.fa-hard-of-hearing:before,.fa-deaf:before{content:""}.fa-glide:before{content:""}.fa-glide-g:before{content:""}.fa-signing:before,.fa-sign-language:before{content:""}.fa-low-vision:before{content:""}.fa-viadeo:before{content:""}.fa-viadeo-square:before{content:""}.fa-snapchat:before{content:""}.fa-snapchat-ghost:before{content:""}.fa-snapchat-square:before{content:""}.fa-pied-piper:before{content:""}.fa-first-order:before{content:""}.fa-yoast:before{content:""}.fa-themeisle:before{content:""}.fa-google-plus-circle:before,.fa-google-plus-official:before{content:""}.fa-fa:before,.fa-font-awesome:before{content:""}.fa-handshake-o:before{content:""}.fa-envelope-open:before{content:""}.fa-envelope-open-o:before{content:""}.fa-linode:before{content:""}.fa-address-book:before{content:""}.fa-address-book-o:before{content:""}.fa-vcard:before,.fa-address-card:before{content:""}.fa-vcard-o:before,.fa-address-card-o:before{content:""}.fa-user-circle:before{content:""}.fa-user-circle-o:before{content:""}.fa-user-o:before{content:""}.fa-id-badge:before{content:""}.fa-drivers-license:before,.fa-id-card:before{content:""}.fa-drivers-license-o:before,.fa-id-card-o:before{content:""}.fa-quora:before{content:""}.fa-free-code-camp:before{content:""}.fa-telegram:before{content:""}.fa-thermometer-4:before,.fa-thermometer:before,.fa-thermometer-full:before{content:""}.fa-thermometer-3:before,.fa-thermometer-three-quarters:before{content:""}.fa-thermometer-2:before,.fa-thermometer-half:before{content:""}.fa-thermometer-1:before,.fa-thermometer-quarter:before{content:""}.fa-thermometer-0:before,.fa-thermometer-empty:before{content:""}.fa-shower:before{content:""}.fa-bathtub:before,.fa-s15:before,.fa-bath:before{content:""}.fa-podcast:before{content:""}.fa-window-maximize:before{content:""}.fa-window-minimize:before{content:""}.fa-window-restore:before{content:""}.fa-times-rectangle:before,.fa-window-close:before{content:""}.fa-times-rectangle-o:before,.fa-window-close-o:before{content:""}.fa-bandcamp:before{content:""}.fa-grav:before{content:""}.fa-etsy:before{content:""}.fa-imdb:before{content:""}.fa-ravelry:before{content:""}.fa-eercast:before{content:""}.fa-microchip:before{content:""}.fa-snowflake-o:before{content:""}.fa-superpowers:before{content:""}.fa-wpexplorer:before{content:""}.fa-meetup:before{content:""}.sr-only{position:absolute;width:1px;height:1px;padding:0;margin:-1px;overflow:hidden;clip:rect(0, 0, 0, 0);border:0}.sr-only-focusable:active,.sr-only-focusable:focus{position:static;width:auto;height:auto;margin:0;overflow:visible;clip:auto}.fa,.wy-menu-vertical li span.toctree-expand,.wy-menu-vertical li.on a span.toctree-expand,.wy-menu-vertical li.current>a span.toctree-expand,.rst-content .admonition-title,.rst-content h1 .headerlink,.rst-content h2 .headerlink,.rst-content h3 .headerlink,.rst-content h4 .headerlink,.rst-content h5 .headerlink,.rst-content h6 .headerlink,.rst-content dl dt .headerlink,.rst-content p.caption .headerlink,.rst-content table>caption .headerlink,.rst-content tt.download span:first-child,.rst-content code.download span:first-child,.icon,.wy-dropdown .caret,.wy-inline-validate.wy-inline-validate-success .wy-input-context,.wy-inline-validate.wy-inline-validate-danger .wy-input-context,.wy-inline-validate.wy-inline-validate-warning .wy-input-context,.wy-inline-validate.wy-inline-validate-info .wy-input-context{font-family:inherit}.fa:before,.wy-menu-vertical li span.toctree-expand:before,.wy-menu-vertical li.on a span.toctree-expand:before,.wy-menu-vertical li.current>a span.toctree-expand:before,.rst-content .admonition-title:before,.rst-content h1 .headerlink:before,.rst-content h2 .headerlink:before,.rst-content h3 .headerlink:before,.rst-content h4 .headerlink:before,.rst-content h5 .headerlink:before,.rst-content h6 .headerlink:before,.rst-content dl dt .headerlink:before,.rst-content p.caption .headerlink:before,.rst-content table>caption .headerlink:before,.rst-content tt.download span:first-child:before,.rst-content code.download span:first-child:before,.icon:before,.wy-dropdown .caret:before,.wy-inline-validate.wy-inline-validate-success .wy-input-context:before,.wy-inline-validate.wy-inline-validate-danger .wy-input-context:before,.wy-inline-validate.wy-inline-validate-warning .wy-input-context:before,.wy-inline-validate.wy-inline-validate-info .wy-input-context:before{font-family:"FontAwesome";display:inline-block;font-style:normal;font-weight:normal;line-height:1;text-decoration:inherit}a .fa,a .wy-menu-vertical li span.toctree-expand,.wy-menu-vertical li a span.toctree-expand,.wy-menu-vertical li.on a span.toctree-expand,.wy-menu-vertical li.current>a span.toctree-expand,a .rst-content .admonition-title,.rst-content a .admonition-title,a .rst-content h1 .headerlink,.rst-content h1 a .headerlink,a .rst-content h2 .headerlink,.rst-content h2 a .headerlink,a .rst-content h3 .headerlink,.rst-content h3 a .headerlink,a .rst-content h4 .headerlink,.rst-content h4 a .headerlink,a .rst-content h5 .headerlink,.rst-content h5 a .headerlink,a .rst-content h6 .headerlink,.rst-content h6 a .headerlink,a .rst-content dl dt .headerlink,.rst-content dl dt a .headerlink,a .rst-content p.caption .headerlink,.rst-content p.caption a .headerlink,a .rst-content table>caption .headerlink,.rst-content table>caption a .headerlink,a .rst-content tt.download span:first-child,.rst-content tt.download a span:first-child,a .rst-content code.download span:first-child,.rst-content code.download a span:first-child,a .icon{display:inline-block;text-decoration:inherit}.btn .fa,.btn .wy-menu-vertical li span.toctree-expand,.wy-menu-vertical li .btn span.toctree-expand,.btn .wy-menu-vertical li.on a span.toctree-expand,.wy-menu-vertical li.on a .btn span.toctree-expand,.btn .wy-menu-vertical li.current>a span.toctree-expand,.wy-menu-vertical li.current>a .btn span.toctree-expand,.btn .rst-content .admonition-title,.rst-content .btn .admonition-title,.btn .rst-content h1 .headerlink,.rst-content h1 .btn .headerlink,.btn .rst-content h2 .headerlink,.rst-content h2 .btn .headerlink,.btn .rst-content h3 .headerlink,.rst-content h3 .btn .headerlink,.btn .rst-content h4 .headerlink,.rst-content h4 .btn .headerlink,.btn .rst-content h5 .headerlink,.rst-content h5 .btn .headerlink,.btn .rst-content h6 .headerlink,.rst-content h6 .btn .headerlink,.btn .rst-content dl dt .headerlink,.rst-content dl dt .btn .headerlink,.btn .rst-content p.caption .headerlink,.rst-content p.caption .btn .headerlink,.btn .rst-content table>caption .headerlink,.rst-content table>caption .btn .headerlink,.btn .rst-content tt.download span:first-child,.rst-content tt.download .btn span:first-child,.btn .rst-content code.download span:first-child,.rst-content code.download .btn span:first-child,.btn .icon,.nav .fa,.nav .wy-menu-vertical li span.toctree-expand,.wy-menu-vertical li .nav span.toctree-expand,.nav .wy-menu-vertical li.on a span.toctree-expand,.wy-menu-vertical li.on a .nav span.toctree-expand,.nav .wy-menu-vertical li.current>a span.toctree-expand,.wy-menu-vertical li.current>a .nav span.toctree-expand,.nav .rst-content .admonition-title,.rst-content .nav .admonition-title,.nav .rst-content h1 .headerlink,.rst-content h1 .nav .headerlink,.nav .rst-content h2 .headerlink,.rst-content h2 .nav .headerlink,.nav .rst-content h3 .headerlink,.rst-content h3 .nav .headerlink,.nav .rst-content h4 .headerlink,.rst-content h4 .nav .headerlink,.nav .rst-content h5 .headerlink,.rst-content h5 .nav .headerlink,.nav .rst-content h6 .headerlink,.rst-content h6 .nav .headerlink,.nav .rst-content dl dt .headerlink,.rst-content dl dt .nav .headerlink,.nav .rst-content p.caption .headerlink,.rst-content p.caption .nav .headerlink,.nav .rst-content table>caption .headerlink,.rst-content table>caption .nav .headerlink,.nav .rst-content tt.download span:first-child,.rst-content tt.download .nav span:first-child,.nav .rst-content code.download span:first-child,.rst-content code.download .nav span:first-child,.nav .icon{display:inline}.btn .fa.fa-large,.btn .wy-menu-vertical li span.fa-large.toctree-expand,.wy-menu-vertical li .btn span.fa-large.toctree-expand,.btn .rst-content .fa-large.admonition-title,.rst-content .btn .fa-large.admonition-title,.btn .rst-content h1 .fa-large.headerlink,.rst-content h1 .btn .fa-large.headerlink,.btn .rst-content h2 .fa-large.headerlink,.rst-content h2 .btn .fa-large.headerlink,.btn .rst-content h3 .fa-large.headerlink,.rst-content h3 .btn .fa-large.headerlink,.btn .rst-content h4 .fa-large.headerlink,.rst-content h4 .btn .fa-large.headerlink,.btn .rst-content h5 .fa-large.headerlink,.rst-content h5 .btn .fa-large.headerlink,.btn .rst-content h6 .fa-large.headerlink,.rst-content h6 .btn .fa-large.headerlink,.btn .rst-content dl dt .fa-large.headerlink,.rst-content dl dt .btn .fa-large.headerlink,.btn .rst-content p.caption .fa-large.headerlink,.rst-content p.caption .btn .fa-large.headerlink,.btn .rst-content table>caption .fa-large.headerlink,.rst-content table>caption .btn .fa-large.headerlink,.btn .rst-content tt.download span.fa-large:first-child,.rst-content tt.download .btn span.fa-large:first-child,.btn .rst-content code.download span.fa-large:first-child,.rst-content code.download .btn span.fa-large:first-child,.btn .fa-large.icon,.nav .fa.fa-large,.nav .wy-menu-vertical li span.fa-large.toctree-expand,.wy-menu-vertical li .nav span.fa-large.toctree-expand,.nav .rst-content .fa-large.admonition-title,.rst-content .nav .fa-large.admonition-title,.nav .rst-content h1 .fa-large.headerlink,.rst-content h1 .nav .fa-large.headerlink,.nav .rst-content h2 .fa-large.headerlink,.rst-content h2 .nav .fa-large.headerlink,.nav .rst-content h3 .fa-large.headerlink,.rst-content h3 .nav .fa-large.headerlink,.nav .rst-content h4 .fa-large.headerlink,.rst-content h4 .nav .fa-large.headerlink,.nav .rst-content h5 .fa-large.headerlink,.rst-content h5 .nav .fa-large.headerlink,.nav .rst-content h6 .fa-large.headerlink,.rst-content h6 .nav .fa-large.headerlink,.nav .rst-content dl dt .fa-large.headerlink,.rst-content dl dt .nav .fa-large.headerlink,.nav .rst-content p.caption .fa-large.headerlink,.rst-content p.caption .nav .fa-large.headerlink,.nav .rst-content table>caption .fa-large.headerlink,.rst-content table>caption .nav .fa-large.headerlink,.nav .rst-content tt.download span.fa-large:first-child,.rst-content tt.download .nav span.fa-large:first-child,.nav .rst-content code.download span.fa-large:first-child,.rst-content code.download .nav span.fa-large:first-child,.nav .fa-large.icon{line-height:.9em}.btn .fa.fa-spin,.btn .wy-menu-vertical li span.fa-spin.toctree-expand,.wy-menu-vertical li .btn span.fa-spin.toctree-expand,.btn .rst-content .fa-spin.admonition-title,.rst-content .btn .fa-spin.admonition-title,.btn .rst-content h1 .fa-spin.headerlink,.rst-content h1 .btn .fa-spin.headerlink,.btn .rst-content h2 .fa-spin.headerlink,.rst-content h2 .btn .fa-spin.headerlink,.btn .rst-content h3 .fa-spin.headerlink,.rst-content h3 .btn .fa-spin.headerlink,.btn .rst-content h4 .fa-spin.headerlink,.rst-content h4 .btn .fa-spin.headerlink,.btn .rst-content h5 .fa-spin.headerlink,.rst-content h5 .btn .fa-spin.headerlink,.btn .rst-content h6 .fa-spin.headerlink,.rst-content h6 .btn .fa-spin.headerlink,.btn .rst-content dl dt .fa-spin.headerlink,.rst-content dl dt .btn .fa-spin.headerlink,.btn .rst-content p.caption .fa-spin.headerlink,.rst-content p.caption .btn .fa-spin.headerlink,.btn .rst-content table>caption .fa-spin.headerlink,.rst-content table>caption .btn .fa-spin.headerlink,.btn .rst-content tt.download span.fa-spin:first-child,.rst-content tt.download .btn span.fa-spin:first-child,.btn .rst-content code.download span.fa-spin:first-child,.rst-content code.download .btn span.fa-spin:first-child,.btn .fa-spin.icon,.nav .fa.fa-spin,.nav .wy-menu-vertical li span.fa-spin.toctree-expand,.wy-menu-vertical li .nav span.fa-spin.toctree-expand,.nav .rst-content .fa-spin.admonition-title,.rst-content .nav .fa-spin.admonition-title,.nav .rst-content h1 .fa-spin.headerlink,.rst-content h1 .nav .fa-spin.headerlink,.nav .rst-content h2 .fa-spin.headerlink,.rst-content h2 .nav .fa-spin.headerlink,.nav .rst-content h3 .fa-spin.headerlink,.rst-content h3 .nav .fa-spin.headerlink,.nav .rst-content h4 .fa-spin.headerlink,.rst-content h4 .nav .fa-spin.headerlink,.nav .rst-content h5 .fa-spin.headerlink,.rst-content h5 .nav .fa-spin.headerlink,.nav .rst-content h6 .fa-spin.headerlink,.rst-content h6 .nav .fa-spin.headerlink,.nav .rst-content dl dt .fa-spin.headerlink,.rst-content dl dt .nav .fa-spin.headerlink,.nav .rst-content p.caption .fa-spin.headerlink,.rst-content p.caption .nav .fa-spin.headerlink,.nav .rst-content table>caption .fa-spin.headerlink,.rst-content table>caption .nav .fa-spin.headerlink,.nav .rst-content tt.download span.fa-spin:first-child,.rst-content tt.download .nav span.fa-spin:first-child,.nav .rst-content code.download span.fa-spin:first-child,.rst-content code.download .nav span.fa-spin:first-child,.nav .fa-spin.icon{display:inline-block}.btn.fa:before,.wy-menu-vertical li span.btn.toctree-expand:before,.rst-content .btn.admonition-title:before,.rst-content h1 .btn.headerlink:before,.rst-content h2 .btn.headerlink:before,.rst-content h3 .btn.headerlink:before,.rst-content h4 .btn.headerlink:before,.rst-content h5 .btn.headerlink:before,.rst-content h6 .btn.headerlink:before,.rst-content dl dt .btn.headerlink:before,.rst-content p.caption .btn.headerlink:before,.rst-content table>caption .btn.headerlink:before,.rst-content tt.download span.btn:first-child:before,.rst-content code.download span.btn:first-child:before,.btn.icon:before{opacity:.5;-webkit-transition:opacity .05s ease-in;-moz-transition:opacity .05s ease-in;transition:opacity .05s ease-in}.btn.fa:hover:before,.wy-menu-vertical li span.btn.toctree-expand:hover:before,.rst-content .btn.admonition-title:hover:before,.rst-content h1 .btn.headerlink:hover:before,.rst-content h2 .btn.headerlink:hover:before,.rst-content h3 .btn.headerlink:hover:before,.rst-content h4 .btn.headerlink:hover:before,.rst-content h5 .btn.headerlink:hover:before,.rst-content h6 .btn.headerlink:hover:before,.rst-content dl dt .btn.headerlink:hover:before,.rst-content p.caption .btn.headerlink:hover:before,.rst-content table>caption .btn.headerlink:hover:before,.rst-content tt.download span.btn:first-child:hover:before,.rst-content code.download span.btn:first-child:hover:before,.btn.icon:hover:before{opacity:1}.btn-mini .fa:before,.btn-mini .wy-menu-vertical li span.toctree-expand:before,.wy-menu-vertical li .btn-mini span.toctree-expand:before,.btn-mini .rst-content .admonition-title:before,.rst-content .btn-mini .admonition-title:before,.btn-mini .rst-content h1 .headerlink:before,.rst-content h1 .btn-mini .headerlink:before,.btn-mini .rst-content h2 .headerlink:before,.rst-content h2 .btn-mini .headerlink:before,.btn-mini .rst-content h3 .headerlink:before,.rst-content h3 .btn-mini .headerlink:before,.btn-mini .rst-content h4 .headerlink:before,.rst-content h4 .btn-mini .headerlink:before,.btn-mini .rst-content h5 .headerlink:before,.rst-content h5 .btn-mini .headerlink:before,.btn-mini .rst-content h6 .headerlink:before,.rst-content h6 .btn-mini .headerlink:before,.btn-mini .rst-content dl dt .headerlink:before,.rst-content dl dt .btn-mini .headerlink:before,.btn-mini .rst-content p.caption .headerlink:before,.rst-content p.caption .btn-mini .headerlink:before,.btn-mini .rst-content table>caption .headerlink:before,.rst-content table>caption .btn-mini .headerlink:before,.btn-mini .rst-content tt.download span:first-child:before,.rst-content tt.download .btn-mini span:first-child:before,.btn-mini .rst-content code.download span:first-child:before,.rst-content code.download .btn-mini span:first-child:before,.btn-mini .icon:before{font-size:14px;vertical-align:-15%}.wy-alert,.rst-content .note,.rst-content .attention,.rst-content .caution,.rst-content .danger,.rst-content .error,.rst-content .hint,.rst-content .important,.rst-content .tip,.rst-content .warning,.rst-content .seealso,.rst-content .admonition-todo,.rst-content .admonition{padding:12px;line-height:24px;margin-bottom:24px;background:#e7f2fa}.wy-alert-title,.rst-content .admonition-title{color:#fff;font-weight:bold;display:block;color:#fff;background:#6ab0de;margin:-12px;padding:6px 12px;margin-bottom:12px}.wy-alert.wy-alert-danger,.rst-content .wy-alert-danger.note,.rst-content .wy-alert-danger.attention,.rst-content .wy-alert-danger.caution,.rst-content .danger,.rst-content .error,.rst-content .wy-alert-danger.hint,.rst-content .wy-alert-danger.important,.rst-content .wy-alert-danger.tip,.rst-content .wy-alert-danger.warning,.rst-content .wy-alert-danger.seealso,.rst-content .wy-alert-danger.admonition-todo,.rst-content .wy-alert-danger.admonition{background:#fdf3f2}.wy-alert.wy-alert-danger .wy-alert-title,.rst-content .wy-alert-danger.note .wy-alert-title,.rst-content .wy-alert-danger.attention .wy-alert-title,.rst-content .wy-alert-danger.caution .wy-alert-title,.rst-content .danger .wy-alert-title,.rst-content .error .wy-alert-title,.rst-content .wy-alert-danger.hint .wy-alert-title,.rst-content .wy-alert-danger.important .wy-alert-title,.rst-content .wy-alert-danger.tip .wy-alert-title,.rst-content .wy-alert-danger.warning .wy-alert-title,.rst-content .wy-alert-danger.seealso .wy-alert-title,.rst-content .wy-alert-danger.admonition-todo .wy-alert-title,.rst-content .wy-alert-danger.admonition .wy-alert-title,.wy-alert.wy-alert-danger .rst-content .admonition-title,.rst-content .wy-alert.wy-alert-danger .admonition-title,.rst-content .wy-alert-danger.note .admonition-title,.rst-content .wy-alert-danger.attention .admonition-title,.rst-content .wy-alert-danger.caution .admonition-title,.rst-content .danger .admonition-title,.rst-content .error .admonition-title,.rst-content .wy-alert-danger.hint .admonition-title,.rst-content .wy-alert-danger.important .admonition-title,.rst-content .wy-alert-danger.tip .admonition-title,.rst-content .wy-alert-danger.warning .admonition-title,.rst-content .wy-alert-danger.seealso .admonition-title,.rst-content .wy-alert-danger.admonition-todo .admonition-title,.rst-content .wy-alert-danger.admonition .admonition-title{background:#f29f97}.wy-alert.wy-alert-warning,.rst-content .wy-alert-warning.note,.rst-content .attention,.rst-content .caution,.rst-content .wy-alert-warning.danger,.rst-content .wy-alert-warning.error,.rst-content .wy-alert-warning.hint,.rst-content .wy-alert-warning.important,.rst-content .wy-alert-warning.tip,.rst-content .warning,.rst-content .wy-alert-warning.seealso,.rst-content .admonition-todo,.rst-content .wy-alert-warning.admonition{background:#ffedcc}.wy-alert.wy-alert-warning .wy-alert-title,.rst-content .wy-alert-warning.note .wy-alert-title,.rst-content .attention .wy-alert-title,.rst-content .caution .wy-alert-title,.rst-content .wy-alert-warning.danger .wy-alert-title,.rst-content .wy-alert-warning.error .wy-alert-title,.rst-content .wy-alert-warning.hint .wy-alert-title,.rst-content .wy-alert-warning.important .wy-alert-title,.rst-content .wy-alert-warning.tip .wy-alert-title,.rst-content .warning .wy-alert-title,.rst-content .wy-alert-warning.seealso .wy-alert-title,.rst-content .admonition-todo .wy-alert-title,.rst-content .wy-alert-warning.admonition .wy-alert-title,.wy-alert.wy-alert-warning .rst-content .admonition-title,.rst-content .wy-alert.wy-alert-warning .admonition-title,.rst-content .wy-alert-warning.note .admonition-title,.rst-content .attention .admonition-title,.rst-content .caution .admonition-title,.rst-content .wy-alert-warning.danger .admonition-title,.rst-content .wy-alert-warning.error .admonition-title,.rst-content .wy-alert-warning.hint .admonition-title,.rst-content .wy-alert-warning.important .admonition-title,.rst-content .wy-alert-warning.tip .admonition-title,.rst-content .warning .admonition-title,.rst-content .wy-alert-warning.seealso .admonition-title,.rst-content .admonition-todo .admonition-title,.rst-content .wy-alert-warning.admonition .admonition-title{background:#f0b37e}.wy-alert.wy-alert-info,.rst-content .note,.rst-content .wy-alert-info.attention,.rst-content .wy-alert-info.caution,.rst-content .wy-alert-info.danger,.rst-content .wy-alert-info.error,.rst-content .wy-alert-info.hint,.rst-content .wy-alert-info.important,.rst-content .wy-alert-info.tip,.rst-content .wy-alert-info.warning,.rst-content .seealso,.rst-content .wy-alert-info.admonition-todo,.rst-content .wy-alert-info.admonition{background:#e7f2fa}.wy-alert.wy-alert-info .wy-alert-title,.rst-content .note .wy-alert-title,.rst-content .wy-alert-info.attention .wy-alert-title,.rst-content .wy-alert-info.caution .wy-alert-title,.rst-content .wy-alert-info.danger .wy-alert-title,.rst-content .wy-alert-info.error .wy-alert-title,.rst-content .wy-alert-info.hint .wy-alert-title,.rst-content .wy-alert-info.important .wy-alert-title,.rst-content .wy-alert-info.tip .wy-alert-title,.rst-content .wy-alert-info.warning .wy-alert-title,.rst-content .seealso .wy-alert-title,.rst-content .wy-alert-info.admonition-todo .wy-alert-title,.rst-content .wy-alert-info.admonition .wy-alert-title,.wy-alert.wy-alert-info .rst-content .admonition-title,.rst-content .wy-alert.wy-alert-info .admonition-title,.rst-content .note .admonition-title,.rst-content .wy-alert-info.attention .admonition-title,.rst-content .wy-alert-info.caution .admonition-title,.rst-content .wy-alert-info.danger .admonition-title,.rst-content .wy-alert-info.error .admonition-title,.rst-content .wy-alert-info.hint .admonition-title,.rst-content .wy-alert-info.important .admonition-title,.rst-content .wy-alert-info.tip .admonition-title,.rst-content .wy-alert-info.warning .admonition-title,.rst-content .seealso .admonition-title,.rst-content .wy-alert-info.admonition-todo .admonition-title,.rst-content .wy-alert-info.admonition .admonition-title{background:#6ab0de}.wy-alert.wy-alert-success,.rst-content .wy-alert-success.note,.rst-content .wy-alert-success.attention,.rst-content .wy-alert-success.caution,.rst-content .wy-alert-success.danger,.rst-content .wy-alert-success.error,.rst-content .hint,.rst-content .important,.rst-content .tip,.rst-content .wy-alert-success.warning,.rst-content .wy-alert-success.seealso,.rst-content .wy-alert-success.admonition-todo,.rst-content .wy-alert-success.admonition{background:#dbfaf4}.wy-alert.wy-alert-success .wy-alert-title,.rst-content .wy-alert-success.note .wy-alert-title,.rst-content .wy-alert-success.attention .wy-alert-title,.rst-content .wy-alert-success.caution .wy-alert-title,.rst-content .wy-alert-success.danger .wy-alert-title,.rst-content .wy-alert-success.error .wy-alert-title,.rst-content .hint .wy-alert-title,.rst-content .important .wy-alert-title,.rst-content .tip .wy-alert-title,.rst-content .wy-alert-success.warning .wy-alert-title,.rst-content .wy-alert-success.seealso .wy-alert-title,.rst-content .wy-alert-success.admonition-todo .wy-alert-title,.rst-content .wy-alert-success.admonition .wy-alert-title,.wy-alert.wy-alert-success .rst-content .admonition-title,.rst-content .wy-alert.wy-alert-success .admonition-title,.rst-content .wy-alert-success.note .admonition-title,.rst-content .wy-alert-success.attention .admonition-title,.rst-content .wy-alert-success.caution .admonition-title,.rst-content .wy-alert-success.danger .admonition-title,.rst-content .wy-alert-success.error .admonition-title,.rst-content .hint .admonition-title,.rst-content .important .admonition-title,.rst-content .tip .admonition-title,.rst-content .wy-alert-success.warning .admonition-title,.rst-content .wy-alert-success.seealso .admonition-title,.rst-content .wy-alert-success.admonition-todo .admonition-title,.rst-content .wy-alert-success.admonition .admonition-title{background:#1abc9c}.wy-alert.wy-alert-neutral,.rst-content .wy-alert-neutral.note,.rst-content .wy-alert-neutral.attention,.rst-content .wy-alert-neutral.caution,.rst-content .wy-alert-neutral.danger,.rst-content .wy-alert-neutral.error,.rst-content .wy-alert-neutral.hint,.rst-content .wy-alert-neutral.important,.rst-content .wy-alert-neutral.tip,.rst-content .wy-alert-neutral.warning,.rst-content .wy-alert-neutral.seealso,.rst-content .wy-alert-neutral.admonition-todo,.rst-content .wy-alert-neutral.admonition{background:#f3f6f6}.wy-alert.wy-alert-neutral .wy-alert-title,.rst-content .wy-alert-neutral.note .wy-alert-title,.rst-content .wy-alert-neutral.attention .wy-alert-title,.rst-content .wy-alert-neutral.caution .wy-alert-title,.rst-content .wy-alert-neutral.danger .wy-alert-title,.rst-content .wy-alert-neutral.error .wy-alert-title,.rst-content .wy-alert-neutral.hint .wy-alert-title,.rst-content .wy-alert-neutral.important .wy-alert-title,.rst-content .wy-alert-neutral.tip .wy-alert-title,.rst-content .wy-alert-neutral.warning .wy-alert-title,.rst-content .wy-alert-neutral.seealso .wy-alert-title,.rst-content .wy-alert-neutral.admonition-todo .wy-alert-title,.rst-content .wy-alert-neutral.admonition .wy-alert-title,.wy-alert.wy-alert-neutral .rst-content .admonition-title,.rst-content .wy-alert.wy-alert-neutral .admonition-title,.rst-content .wy-alert-neutral.note .admonition-title,.rst-content .wy-alert-neutral.attention .admonition-title,.rst-content .wy-alert-neutral.caution .admonition-title,.rst-content .wy-alert-neutral.danger .admonition-title,.rst-content .wy-alert-neutral.error .admonition-title,.rst-content .wy-alert-neutral.hint .admonition-title,.rst-content .wy-alert-neutral.important .admonition-title,.rst-content .wy-alert-neutral.tip .admonition-title,.rst-content .wy-alert-neutral.warning .admonition-title,.rst-content .wy-alert-neutral.seealso .admonition-title,.rst-content .wy-alert-neutral.admonition-todo .admonition-title,.rst-content .wy-alert-neutral.admonition .admonition-title{color:#404040;background:#e1e4e5}.wy-alert.wy-alert-neutral a,.rst-content .wy-alert-neutral.note a,.rst-content .wy-alert-neutral.attention a,.rst-content .wy-alert-neutral.caution a,.rst-content .wy-alert-neutral.danger a,.rst-content .wy-alert-neutral.error a,.rst-content .wy-alert-neutral.hint a,.rst-content .wy-alert-neutral.important a,.rst-content .wy-alert-neutral.tip a,.rst-content .wy-alert-neutral.warning a,.rst-content .wy-alert-neutral.seealso a,.rst-content .wy-alert-neutral.admonition-todo a,.rst-content .wy-alert-neutral.admonition a{color:#2980B9}.wy-alert p:last-child,.rst-content .note p:last-child,.rst-content .attention p:last-child,.rst-content .caution p:last-child,.rst-content .danger p:last-child,.rst-content .error p:last-child,.rst-content .hint p:last-child,.rst-content .important p:last-child,.rst-content .tip p:last-child,.rst-content .warning p:last-child,.rst-content .seealso p:last-child,.rst-content .admonition-todo p:last-child,.rst-content .admonition p:last-child{margin-bottom:0}.wy-tray-container{position:fixed;bottom:0px;left:0;z-index:600}.wy-tray-container li{display:block;width:300px;background:transparent;color:#fff;text-align:center;box-shadow:0 5px 5px 0 rgba(0,0,0,0.1);padding:0 24px;min-width:20%;opacity:0;height:0;line-height:56px;overflow:hidden;-webkit-transition:all .3s ease-in;-moz-transition:all .3s ease-in;transition:all .3s ease-in}.wy-tray-container li.wy-tray-item-success{background:#27AE60}.wy-tray-container li.wy-tray-item-info{background:#2980B9}.wy-tray-container li.wy-tray-item-warning{background:#E67E22}.wy-tray-container li.wy-tray-item-danger{background:#E74C3C}.wy-tray-container li.on{opacity:1;height:56px}@media screen and (max-width: 768px){.wy-tray-container{bottom:auto;top:0;width:100%}.wy-tray-container li{width:100%}}button{font-size:100%;margin:0;vertical-align:baseline;*vertical-align:middle;cursor:pointer;line-height:normal;-webkit-appearance:button;*overflow:visible}button::-moz-focus-inner,input::-moz-focus-inner{border:0;padding:0}button[disabled]{cursor:default}.btn{display:inline-block;border-radius:2px;line-height:normal;white-space:nowrap;text-align:center;cursor:pointer;font-size:100%;padding:6px 12px 8px 12px;color:#fff;border:1px solid rgba(0,0,0,0.1);background-color:#27AE60;text-decoration:none;font-weight:normal;font-family:"Lato","proxima-nova","Helvetica Neue",Arial,sans-serif;box-shadow:0px 1px 2px -1px rgba(255,255,255,0.5) inset,0px -2px 0px 0px rgba(0,0,0,0.1) inset;outline-none:false;vertical-align:middle;*display:inline;zoom:1;-webkit-user-drag:none;-webkit-user-select:none;-moz-user-select:none;-ms-user-select:none;user-select:none;-webkit-transition:all .1s linear;-moz-transition:all .1s linear;transition:all .1s linear}.btn-hover{background:#2e8ece;color:#fff}.btn:hover{background:#2cc36b;color:#fff}.btn:focus{background:#2cc36b;outline:0}.btn:active{box-shadow:0px -1px 0px 0px rgba(0,0,0,0.05) inset,0px 2px 0px 0px rgba(0,0,0,0.1) inset;padding:8px 12px 6px 12px}.btn:visited{color:#fff}.btn:disabled{background-image:none;filter:progid:DXImageTransform.Microsoft.gradient(enabled = false);filter:alpha(opacity=40);opacity:.4;cursor:not-allowed;box-shadow:none}.btn-disabled{background-image:none;filter:progid:DXImageTransform.Microsoft.gradient(enabled = false);filter:alpha(opacity=40);opacity:.4;cursor:not-allowed;box-shadow:none}.btn-disabled:hover,.btn-disabled:focus,.btn-disabled:active{background-image:none;filter:progid:DXImageTransform.Microsoft.gradient(enabled = false);filter:alpha(opacity=40);opacity:.4;cursor:not-allowed;box-shadow:none}.btn::-moz-focus-inner{padding:0;border:0}.btn-small{font-size:80%}.btn-info{background-color:#2980B9 !important}.btn-info:hover{background-color:#2e8ece !important}.btn-neutral{background-color:#f3f6f6 !important;color:#404040 !important}.btn-neutral:hover{background-color:#e5ebeb !important;color:#404040}.btn-neutral:visited{color:#404040 !important}.btn-success{background-color:#27AE60 !important}.btn-success:hover{background-color:#295 !important}.btn-danger{background-color:#E74C3C !important}.btn-danger:hover{background-color:#ea6153 !important}.btn-warning{background-color:#E67E22 !important}.btn-warning:hover{background-color:#e98b39 !important}.btn-invert{background-color:#222}.btn-invert:hover{background-color:#2f2f2f !important}.btn-link{background-color:transparent !important;color:#2980B9;box-shadow:none;border-color:transparent !important}.btn-link:hover{background-color:transparent !important;color:#409ad5 !important;box-shadow:none}.btn-link:active{background-color:transparent !important;color:#409ad5 !important;box-shadow:none}.btn-link:visited{color:#9B59B6}.wy-btn-group .btn,.wy-control .btn{vertical-align:middle}.wy-btn-group{margin-bottom:24px;*zoom:1}.wy-btn-group:before,.wy-btn-group:after{display:table;content:""}.wy-btn-group:after{clear:both}.wy-dropdown{position:relative;display:inline-block}.wy-dropdown-active .wy-dropdown-menu{display:block}.wy-dropdown-menu{position:absolute;left:0;display:none;float:left;top:100%;min-width:100%;background:#fcfcfc;z-index:100;border:solid 1px #cfd7dd;box-shadow:0 2px 2px 0 rgba(0,0,0,0.1);padding:12px}.wy-dropdown-menu>dd>a{display:block;clear:both;color:#404040;white-space:nowrap;font-size:90%;padding:0 12px;cursor:pointer}.wy-dropdown-menu>dd>a:hover{background:#2980B9;color:#fff}.wy-dropdown-menu>dd.divider{border-top:solid 1px #cfd7dd;margin:6px 0}.wy-dropdown-menu>dd.search{padding-bottom:12px}.wy-dropdown-menu>dd.search input[type="search"]{width:100%}.wy-dropdown-menu>dd.call-to-action{background:#e3e3e3;text-transform:uppercase;font-weight:500;font-size:80%}.wy-dropdown-menu>dd.call-to-action:hover{background:#e3e3e3}.wy-dropdown-menu>dd.call-to-action .btn{color:#fff}.wy-dropdown.wy-dropdown-up .wy-dropdown-menu{bottom:100%;top:auto;left:auto;right:0}.wy-dropdown.wy-dropdown-bubble .wy-dropdown-menu{background:#fcfcfc;margin-top:2px}.wy-dropdown.wy-dropdown-bubble .wy-dropdown-menu a{padding:6px 12px}.wy-dropdown.wy-dropdown-bubble .wy-dropdown-menu a:hover{background:#2980B9;color:#fff}.wy-dropdown.wy-dropdown-left .wy-dropdown-menu{right:0;left:auto;text-align:right}.wy-dropdown-arrow:before{content:" ";border-bottom:5px solid #f5f5f5;border-left:5px solid transparent;border-right:5px solid transparent;position:absolute;display:block;top:-4px;left:50%;margin-left:-3px}.wy-dropdown-arrow.wy-dropdown-arrow-left:before{left:11px}.wy-form-stacked select{display:block}.wy-form-aligned input,.wy-form-aligned textarea,.wy-form-aligned select,.wy-form-aligned .wy-help-inline,.wy-form-aligned label{display:inline-block;*display:inline;*zoom:1;vertical-align:middle}.wy-form-aligned .wy-control-group>label{display:inline-block;vertical-align:middle;width:10em;margin:6px 12px 0 0;float:left}.wy-form-aligned .wy-control{float:left}.wy-form-aligned .wy-control label{display:block}.wy-form-aligned .wy-control select{margin-top:6px}fieldset{border:0;margin:0;padding:0}legend{display:block;width:100%;border:0;padding:0;white-space:normal;margin-bottom:24px;font-size:150%;*margin-left:-7px}label{display:block;margin:0 0 .3125em 0;color:#333;font-size:90%}input,select,textarea{font-size:100%;margin:0;vertical-align:baseline;*vertical-align:middle}.wy-control-group{margin-bottom:24px;*zoom:1;max-width:68em;margin-left:auto;margin-right:auto;*zoom:1}.wy-control-group:before,.wy-control-group:after{display:table;content:""}.wy-control-group:after{clear:both}.wy-control-group:before,.wy-control-group:after{display:table;content:""}.wy-control-group:after{clear:both}.wy-control-group.wy-control-group-required>label:after{content:" *";color:#E74C3C}.wy-control-group .wy-form-full,.wy-control-group .wy-form-halves,.wy-control-group .wy-form-thirds{padding-bottom:12px}.wy-control-group .wy-form-full select,.wy-control-group .wy-form-halves select,.wy-control-group .wy-form-thirds select{width:100%}.wy-control-group .wy-form-full input[type="text"],.wy-control-group .wy-form-full input[type="password"],.wy-control-group .wy-form-full input[type="email"],.wy-control-group .wy-form-full input[type="url"],.wy-control-group .wy-form-full input[type="date"],.wy-control-group .wy-form-full input[type="month"],.wy-control-group .wy-form-full input[type="time"],.wy-control-group .wy-form-full input[type="datetime"],.wy-control-group .wy-form-full input[type="datetime-local"],.wy-control-group .wy-form-full input[type="week"],.wy-control-group .wy-form-full input[type="number"],.wy-control-group .wy-form-full input[type="search"],.wy-control-group .wy-form-full input[type="tel"],.wy-control-group .wy-form-full input[type="color"],.wy-control-group .wy-form-halves input[type="text"],.wy-control-group .wy-form-halves input[type="password"],.wy-control-group .wy-form-halves input[type="email"],.wy-control-group .wy-form-halves input[type="url"],.wy-control-group .wy-form-halves input[type="date"],.wy-control-group .wy-form-halves input[type="month"],.wy-control-group .wy-form-halves input[type="time"],.wy-control-group .wy-form-halves input[type="datetime"],.wy-control-group .wy-form-halves input[type="datetime-local"],.wy-control-group .wy-form-halves input[type="week"],.wy-control-group .wy-form-halves input[type="number"],.wy-control-group .wy-form-halves input[type="search"],.wy-control-group .wy-form-halves input[type="tel"],.wy-control-group .wy-form-halves input[type="color"],.wy-control-group .wy-form-thirds input[type="text"],.wy-control-group .wy-form-thirds input[type="password"],.wy-control-group .wy-form-thirds input[type="email"],.wy-control-group .wy-form-thirds input[type="url"],.wy-control-group .wy-form-thirds input[type="date"],.wy-control-group .wy-form-thirds input[type="month"],.wy-control-group .wy-form-thirds input[type="time"],.wy-control-group .wy-form-thirds input[type="datetime"],.wy-control-group .wy-form-thirds input[type="datetime-local"],.wy-control-group .wy-form-thirds input[type="week"],.wy-control-group .wy-form-thirds input[type="number"],.wy-control-group .wy-form-thirds input[type="search"],.wy-control-group .wy-form-thirds input[type="tel"],.wy-control-group .wy-form-thirds input[type="color"]{width:100%}.wy-control-group .wy-form-full{float:left;display:block;margin-right:2.3576515979%;width:100%;margin-right:0}.wy-control-group .wy-form-full:last-child{margin-right:0}.wy-control-group .wy-form-halves{float:left;display:block;margin-right:2.3576515979%;width:48.821174201%}.wy-control-group .wy-form-halves:last-child{margin-right:0}.wy-control-group .wy-form-halves:nth-of-type(2n){margin-right:0}.wy-control-group .wy-form-halves:nth-of-type(2n+1){clear:left}.wy-control-group .wy-form-thirds{float:left;display:block;margin-right:2.3576515979%;width:31.7615656014%}.wy-control-group .wy-form-thirds:last-child{margin-right:0}.wy-control-group .wy-form-thirds:nth-of-type(3n){margin-right:0}.wy-control-group .wy-form-thirds:nth-of-type(3n+1){clear:left}.wy-control-group.wy-control-group-no-input .wy-control{margin:6px 0 0 0;font-size:90%}.wy-control-no-input{display:inline-block;margin:6px 0 0 0;font-size:90%}.wy-control-group.fluid-input input[type="text"],.wy-control-group.fluid-input input[type="password"],.wy-control-group.fluid-input input[type="email"],.wy-control-group.fluid-input input[type="url"],.wy-control-group.fluid-input input[type="date"],.wy-control-group.fluid-input input[type="month"],.wy-control-group.fluid-input input[type="time"],.wy-control-group.fluid-input input[type="datetime"],.wy-control-group.fluid-input input[type="datetime-local"],.wy-control-group.fluid-input input[type="week"],.wy-control-group.fluid-input input[type="number"],.wy-control-group.fluid-input input[type="search"],.wy-control-group.fluid-input input[type="tel"],.wy-control-group.fluid-input input[type="color"]{width:100%}.wy-form-message-inline{display:inline-block;padding-left:.3em;color:#666;vertical-align:middle;font-size:90%}.wy-form-message{display:block;color:#999;font-size:70%;margin-top:.3125em;font-style:italic}.wy-form-message p{font-size:inherit;font-style:italic;margin-bottom:6px}.wy-form-message p:last-child{margin-bottom:0}input{line-height:normal}input[type="button"],input[type="reset"],input[type="submit"]{-webkit-appearance:button;cursor:pointer;font-family:"Lato","proxima-nova","Helvetica Neue",Arial,sans-serif;*overflow:visible}input[type="text"],input[type="password"],input[type="email"],input[type="url"],input[type="date"],input[type="month"],input[type="time"],input[type="datetime"],input[type="datetime-local"],input[type="week"],input[type="number"],input[type="search"],input[type="tel"],input[type="color"]{-webkit-appearance:none;padding:6px;display:inline-block;border:1px solid #ccc;font-size:80%;font-family:"Lato","proxima-nova","Helvetica Neue",Arial,sans-serif;box-shadow:inset 0 1px 3px #ddd;border-radius:0;-webkit-transition:border .3s linear;-moz-transition:border .3s linear;transition:border .3s linear}input[type="datetime-local"]{padding:.34375em .625em}input[disabled]{cursor:default}input[type="checkbox"],input[type="radio"]{-webkit-box-sizing:border-box;-moz-box-sizing:border-box;box-sizing:border-box;padding:0;margin-right:.3125em;*height:13px;*width:13px}input[type="search"]{-webkit-box-sizing:border-box;-moz-box-sizing:border-box;box-sizing:border-box}input[type="search"]::-webkit-search-cancel-button,input[type="search"]::-webkit-search-decoration{-webkit-appearance:none}input[type="text"]:focus,input[type="password"]:focus,input[type="email"]:focus,input[type="url"]:focus,input[type="date"]:focus,input[type="month"]:focus,input[type="time"]:focus,input[type="datetime"]:focus,input[type="datetime-local"]:focus,input[type="week"]:focus,input[type="number"]:focus,input[type="search"]:focus,input[type="tel"]:focus,input[type="color"]:focus{outline:0;outline:thin dotted \9;border-color:#333}input.no-focus:focus{border-color:#ccc !important}input[type="file"]:focus,input[type="radio"]:focus,input[type="checkbox"]:focus{outline:thin dotted #333;outline:1px auto #129FEA}input[type="text"][disabled],input[type="password"][disabled],input[type="email"][disabled],input[type="url"][disabled],input[type="date"][disabled],input[type="month"][disabled],input[type="time"][disabled],input[type="datetime"][disabled],input[type="datetime-local"][disabled],input[type="week"][disabled],input[type="number"][disabled],input[type="search"][disabled],input[type="tel"][disabled],input[type="color"][disabled]{cursor:not-allowed;background-color:#fafafa}input:focus:invalid,textarea:focus:invalid,select:focus:invalid{color:#E74C3C;border:1px solid #E74C3C}input:focus:invalid:focus,textarea:focus:invalid:focus,select:focus:invalid:focus{border-color:#E74C3C}input[type="file"]:focus:invalid:focus,input[type="radio"]:focus:invalid:focus,input[type="checkbox"]:focus:invalid:focus{outline-color:#E74C3C}input.wy-input-large{padding:12px;font-size:100%}textarea{overflow:auto;vertical-align:top;width:100%;font-family:"Lato","proxima-nova","Helvetica Neue",Arial,sans-serif}select,textarea{padding:.5em .625em;display:inline-block;border:1px solid #ccc;font-size:80%;box-shadow:inset 0 1px 3px #ddd;-webkit-transition:border .3s linear;-moz-transition:border .3s linear;transition:border .3s linear}select{border:1px solid #ccc;background-color:#fff}select[multiple]{height:auto}select:focus,textarea:focus{outline:0}select[disabled],textarea[disabled],input[readonly],select[readonly],textarea[readonly]{cursor:not-allowed;background-color:#fafafa}input[type="radio"][disabled],input[type="checkbox"][disabled]{cursor:not-allowed}.wy-checkbox,.wy-radio{margin:6px 0;color:#404040;display:block}.wy-checkbox input,.wy-radio input{vertical-align:baseline}.wy-form-message-inline{display:inline-block;*display:inline;*zoom:1;vertical-align:middle}.wy-input-prefix,.wy-input-suffix{white-space:nowrap;padding:6px}.wy-input-prefix .wy-input-context,.wy-input-suffix .wy-input-context{line-height:27px;padding:0 8px;display:inline-block;font-size:80%;background-color:#f3f6f6;border:solid 1px #ccc;color:#999}.wy-input-suffix .wy-input-context{border-left:0}.wy-input-prefix .wy-input-context{border-right:0}.wy-switch{position:relative;display:block;height:24px;margin-top:12px;cursor:pointer}.wy-switch:before{position:absolute;content:"";display:block;left:0;top:0;width:36px;height:12px;border-radius:4px;background:#ccc;-webkit-transition:all .2s ease-in-out;-moz-transition:all .2s ease-in-out;transition:all .2s ease-in-out}.wy-switch:after{position:absolute;content:"";display:block;width:18px;height:18px;border-radius:4px;background:#999;left:-3px;top:-3px;-webkit-transition:all .2s ease-in-out;-moz-transition:all .2s ease-in-out;transition:all .2s ease-in-out}.wy-switch span{position:absolute;left:48px;display:block;font-size:12px;color:#ccc;line-height:1}.wy-switch.active:before{background:#1e8449}.wy-switch.active:after{left:24px;background:#27AE60}.wy-switch.disabled{cursor:not-allowed;opacity:.8}.wy-control-group.wy-control-group-error .wy-form-message,.wy-control-group.wy-control-group-error>label{color:#E74C3C}.wy-control-group.wy-control-group-error input[type="text"],.wy-control-group.wy-control-group-error input[type="password"],.wy-control-group.wy-control-group-error input[type="email"],.wy-control-group.wy-control-group-error input[type="url"],.wy-control-group.wy-control-group-error input[type="date"],.wy-control-group.wy-control-group-error input[type="month"],.wy-control-group.wy-control-group-error input[type="time"],.wy-control-group.wy-control-group-error input[type="datetime"],.wy-control-group.wy-control-group-error input[type="datetime-local"],.wy-control-group.wy-control-group-error input[type="week"],.wy-control-group.wy-control-group-error input[type="number"],.wy-control-group.wy-control-group-error input[type="search"],.wy-control-group.wy-control-group-error input[type="tel"],.wy-control-group.wy-control-group-error input[type="color"]{border:solid 1px #E74C3C}.wy-control-group.wy-control-group-error textarea{border:solid 1px #E74C3C}.wy-inline-validate{white-space:nowrap}.wy-inline-validate .wy-input-context{padding:.5em .625em;display:inline-block;font-size:80%}.wy-inline-validate.wy-inline-validate-success .wy-input-context{color:#27AE60}.wy-inline-validate.wy-inline-validate-danger .wy-input-context{color:#E74C3C}.wy-inline-validate.wy-inline-validate-warning .wy-input-context{color:#E67E22}.wy-inline-validate.wy-inline-validate-info .wy-input-context{color:#2980B9}.rotate-90{-webkit-transform:rotate(90deg);-moz-transform:rotate(90deg);-ms-transform:rotate(90deg);-o-transform:rotate(90deg);transform:rotate(90deg)}.rotate-180{-webkit-transform:rotate(180deg);-moz-transform:rotate(180deg);-ms-transform:rotate(180deg);-o-transform:rotate(180deg);transform:rotate(180deg)}.rotate-270{-webkit-transform:rotate(270deg);-moz-transform:rotate(270deg);-ms-transform:rotate(270deg);-o-transform:rotate(270deg);transform:rotate(270deg)}.mirror{-webkit-transform:scaleX(-1);-moz-transform:scaleX(-1);-ms-transform:scaleX(-1);-o-transform:scaleX(-1);transform:scaleX(-1)}.mirror.rotate-90{-webkit-transform:scaleX(-1) rotate(90deg);-moz-transform:scaleX(-1) rotate(90deg);-ms-transform:scaleX(-1) rotate(90deg);-o-transform:scaleX(-1) rotate(90deg);transform:scaleX(-1) rotate(90deg)}.mirror.rotate-180{-webkit-transform:scaleX(-1) rotate(180deg);-moz-transform:scaleX(-1) rotate(180deg);-ms-transform:scaleX(-1) rotate(180deg);-o-transform:scaleX(-1) rotate(180deg);transform:scaleX(-1) rotate(180deg)}.mirror.rotate-270{-webkit-transform:scaleX(-1) rotate(270deg);-moz-transform:scaleX(-1) rotate(270deg);-ms-transform:scaleX(-1) rotate(270deg);-o-transform:scaleX(-1) rotate(270deg);transform:scaleX(-1) rotate(270deg)}@media only screen and (max-width: 480px){.wy-form button[type="submit"]{margin:.7em 0 0}.wy-form input[type="text"],.wy-form input[type="password"],.wy-form input[type="email"],.wy-form input[type="url"],.wy-form input[type="date"],.wy-form input[type="month"],.wy-form input[type="time"],.wy-form input[type="datetime"],.wy-form input[type="datetime-local"],.wy-form input[type="week"],.wy-form input[type="number"],.wy-form input[type="search"],.wy-form input[type="tel"],.wy-form input[type="color"]{margin-bottom:.3em;display:block}.wy-form label{margin-bottom:.3em;display:block}.wy-form input[type="password"],.wy-form input[type="email"],.wy-form input[type="url"],.wy-form input[type="date"],.wy-form input[type="month"],.wy-form input[type="time"],.wy-form input[type="datetime"],.wy-form input[type="datetime-local"],.wy-form input[type="week"],.wy-form input[type="number"],.wy-form input[type="search"],.wy-form input[type="tel"],.wy-form input[type="color"]{margin-bottom:0}.wy-form-aligned .wy-control-group label{margin-bottom:.3em;text-align:left;display:block;width:100%}.wy-form-aligned .wy-control{margin:1.5em 0 0 0}.wy-form .wy-help-inline,.wy-form-message-inline,.wy-form-message{display:block;font-size:80%;padding:6px 0}}@media screen and (max-width: 768px){.tablet-hide{display:none}}@media screen and (max-width: 480px){.mobile-hide{display:none}}.float-left{float:left}.float-right{float:right}.full-width{width:100%}.wy-table,.rst-content table.docutils,.rst-content table.field-list{border-collapse:collapse;border-spacing:0;empty-cells:show;margin-bottom:24px}.wy-table caption,.rst-content table.docutils caption,.rst-content table.field-list caption{color:#000;font:italic 85%/1 arial,sans-serif;padding:1em 0;text-align:center}.wy-table td,.rst-content table.docutils td,.rst-content table.field-list td,.wy-table th,.rst-content table.docutils th,.rst-content table.field-list th{font-size:90%;margin:0;overflow:visible;padding:8px 16px}.wy-table td:first-child,.rst-content table.docutils td:first-child,.rst-content table.field-list td:first-child,.wy-table th:first-child,.rst-content table.docutils th:first-child,.rst-content table.field-list th:first-child{border-left-width:0}.wy-table thead,.rst-content table.docutils thead,.rst-content table.field-list thead{color:#000;text-align:left;vertical-align:bottom;white-space:nowrap}.wy-table thead th,.rst-content table.docutils thead th,.rst-content table.field-list thead th{font-weight:bold;border-bottom:solid 2px #e1e4e5}.wy-table td,.rst-content table.docutils td,.rst-content table.field-list td{background-color:transparent;vertical-align:middle}.wy-table td p,.rst-content table.docutils td p,.rst-content table.field-list td p{line-height:18px}.wy-table td p:last-child,.rst-content table.docutils td p:last-child,.rst-content table.field-list td p:last-child{margin-bottom:0}.wy-table .wy-table-cell-min,.rst-content table.docutils .wy-table-cell-min,.rst-content table.field-list .wy-table-cell-min{width:1%;padding-right:0}.wy-table .wy-table-cell-min input[type=checkbox],.rst-content table.docutils .wy-table-cell-min input[type=checkbox],.rst-content table.field-list .wy-table-cell-min input[type=checkbox],.wy-table .wy-table-cell-min input[type=checkbox],.rst-content table.docutils .wy-table-cell-min input[type=checkbox],.rst-content table.field-list .wy-table-cell-min input[type=checkbox]{margin:0}.wy-table-secondary{color:gray;font-size:90%}.wy-table-tertiary{color:gray;font-size:80%}.wy-table-odd td,.wy-table-striped tr:nth-child(2n-1) td,.rst-content table.docutils:not(.field-list) tr:nth-child(2n-1) td{background-color:#f3f6f6}.wy-table-backed{background-color:#f3f6f6}.wy-table-bordered-all,.rst-content table.docutils{border:1px solid #e1e4e5}.wy-table-bordered-all td,.rst-content table.docutils td{border-bottom:1px solid #e1e4e5;border-left:1px solid #e1e4e5}.wy-table-bordered-all tbody>tr:last-child td,.rst-content table.docutils tbody>tr:last-child td{border-bottom-width:0}.wy-table-bordered{border:1px solid #e1e4e5}.wy-table-bordered-rows td{border-bottom:1px solid #e1e4e5}.wy-table-bordered-rows tbody>tr:last-child td{border-bottom-width:0}.wy-table-horizontal tbody>tr:last-child td{border-bottom-width:0}.wy-table-horizontal td,.wy-table-horizontal th{border-width:0 0 1px 0;border-bottom:1px solid #e1e4e5}.wy-table-horizontal tbody>tr:last-child td{border-bottom-width:0}.wy-table-responsive{margin-bottom:24px;max-width:100%;overflow:auto}.wy-table-responsive table{margin-bottom:0 !important}.wy-table-responsive table td,.wy-table-responsive table th{white-space:nowrap}a{color:#2980B9;text-decoration:none;cursor:pointer}a:hover{color:#3091d1}a:visited{color:#9B59B6}html{height:100%;overflow-x:hidden}body{font-family:"Lato","proxima-nova","Helvetica Neue",Arial,sans-serif;font-weight:normal;color:#404040;min-height:100%;overflow-x:hidden;background:#edf0f2}.wy-text-left{text-align:left}.wy-text-center{text-align:center}.wy-text-right{text-align:right}.wy-text-large{font-size:120%}.wy-text-normal{font-size:100%}.wy-text-small,small{font-size:80%}.wy-text-strike{text-decoration:line-through}.wy-text-warning{color:#E67E22 !important}a.wy-text-warning:hover{color:#eb9950 !important}.wy-text-info{color:#2980B9 !important}a.wy-text-info:hover{color:#409ad5 !important}.wy-text-success{color:#27AE60 !important}a.wy-text-success:hover{color:#36d278 !important}.wy-text-danger{color:#E74C3C !important}a.wy-text-danger:hover{color:#ed7669 !important}.wy-text-neutral{color:#404040 !important}a.wy-text-neutral:hover{color:#595959 !important}h1,h2,.rst-content .toctree-wrapper p.caption,h3,h4,h5,h6,legend{margin-top:0;font-weight:700;font-family:"Roboto Slab","ff-tisa-web-pro","Georgia",Arial,sans-serif}p{line-height:24px;margin:0;font-size:16px;margin-bottom:24px}h1{font-size:175%}h2,.rst-content .toctree-wrapper p.caption{font-size:150%}h3{font-size:125%}h4{font-size:115%}h5{font-size:110%}h6{font-size:100%}hr{display:block;height:1px;border:0;border-top:1px solid #e1e4e5;margin:24px 0;padding:0}code,.rst-content tt,.rst-content code{white-space:nowrap;max-width:100%;background:#fff;border:solid 1px #e1e4e5;font-size:75%;padding:0 5px;font-family:SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace;color:#E74C3C;overflow-x:auto}code.code-large,.rst-content tt.code-large{font-size:90%}.wy-plain-list-disc,.rst-content .section ul,.rst-content .toctree-wrapper ul,article ul{list-style:disc;line-height:24px;margin-bottom:24px}.wy-plain-list-disc li,.rst-content .section ul li,.rst-content .toctree-wrapper ul li,article ul li{list-style:disc;margin-left:24px}.wy-plain-list-disc li p:last-child,.rst-content .section ul li p:last-child,.rst-content .toctree-wrapper ul li p:last-child,article ul li p:last-child{margin-bottom:0}.wy-plain-list-disc li ul,.rst-content .section ul li ul,.rst-content .toctree-wrapper ul li ul,article ul li ul{margin-bottom:0}.wy-plain-list-disc li li,.rst-content .section ul li li,.rst-content .toctree-wrapper ul li li,article ul li li{list-style:circle}.wy-plain-list-disc li li li,.rst-content .section ul li li li,.rst-content .toctree-wrapper ul li li li,article ul li li li{list-style:square}.wy-plain-list-disc li ol li,.rst-content .section ul li ol li,.rst-content .toctree-wrapper ul li ol li,article ul li ol li{list-style:decimal}.wy-plain-list-decimal,.rst-content .section ol,.rst-content ol.arabic,article ol{list-style:decimal;line-height:24px;margin-bottom:24px}.wy-plain-list-decimal li,.rst-content .section ol li,.rst-content ol.arabic li,article ol li{list-style:decimal;margin-left:24px}.wy-plain-list-decimal li p:last-child,.rst-content .section ol li p:last-child,.rst-content ol.arabic li p:last-child,article ol li p:last-child{margin-bottom:0}.wy-plain-list-decimal li ul,.rst-content .section ol li ul,.rst-content ol.arabic li ul,article ol li ul{margin-bottom:0}.wy-plain-list-decimal li ul li,.rst-content .section ol li ul li,.rst-content ol.arabic li ul li,article ol li ul li{list-style:disc}.wy-breadcrumbs{*zoom:1}.wy-breadcrumbs:before,.wy-breadcrumbs:after{display:table;content:""}.wy-breadcrumbs:after{clear:both}.wy-breadcrumbs li{display:inline-block}.wy-breadcrumbs li.wy-breadcrumbs-aside{float:right}.wy-breadcrumbs li a{display:inline-block;padding:5px}.wy-breadcrumbs li a:first-child{padding-left:0}.wy-breadcrumbs li code,.wy-breadcrumbs li .rst-content tt,.rst-content .wy-breadcrumbs li tt{padding:5px;border:none;background:none}.wy-breadcrumbs li code.literal,.wy-breadcrumbs li .rst-content tt.literal,.rst-content .wy-breadcrumbs li tt.literal{color:#404040}.wy-breadcrumbs-extra{margin-bottom:0;color:#b3b3b3;font-size:80%;display:inline-block}@media screen and (max-width: 480px){.wy-breadcrumbs-extra{display:none}.wy-breadcrumbs li.wy-breadcrumbs-aside{display:none}}@media print{.wy-breadcrumbs li.wy-breadcrumbs-aside{display:none}}.wy-affix{position:fixed;top:1.618em}.wy-menu a:hover{text-decoration:none}.wy-menu-horiz{*zoom:1}.wy-menu-horiz:before,.wy-menu-horiz:after{display:table;content:""}.wy-menu-horiz:after{clear:both}.wy-menu-horiz ul,.wy-menu-horiz li{display:inline-block}.wy-menu-horiz li:hover{background:rgba(255,255,255,0.1)}.wy-menu-horiz li.divide-left{border-left:solid 1px #404040}.wy-menu-horiz li.divide-right{border-right:solid 1px #404040}.wy-menu-horiz a{height:32px;display:inline-block;line-height:32px;padding:0 16px}.wy-menu-vertical{width:300px}.wy-menu-vertical header,.wy-menu-vertical p.caption{height:32px;display:inline-block;line-height:32px;padding:0 1.618em;margin-bottom:0;display:block;font-weight:bold;text-transform:uppercase;font-size:80%;white-space:nowrap}.wy-menu-vertical ul{margin-bottom:0}.wy-menu-vertical li.divide-top{border-top:solid 1px #404040}.wy-menu-vertical li.divide-bottom{border-bottom:solid 1px #404040}.wy-menu-vertical li.current{background:#e3e3e3}.wy-menu-vertical li.current a{color:gray;border-right:solid 1px #c9c9c9;padding:.4045em 2.427em}.wy-menu-vertical li.current a:hover{background:#d6d6d6}.wy-menu-vertical li code,.wy-menu-vertical li .rst-content tt,.rst-content .wy-menu-vertical li tt{border:none;background:inherit;color:inherit;padding-left:0;padding-right:0}.wy-menu-vertical li span.toctree-expand{display:block;float:left;margin-left:-1.2em;font-size:.8em;line-height:1.6em;color:#4d4d4d}.wy-menu-vertical li.on a,.wy-menu-vertical li.current>a{color:#404040;padding:.4045em 1.618em;font-weight:bold;position:relative;background:#fcfcfc;border:none;padding-left:1.618em -4px}.wy-menu-vertical li.on a:hover,.wy-menu-vertical li.current>a:hover{background:#fcfcfc}.wy-menu-vertical li.on a:hover span.toctree-expand,.wy-menu-vertical li.current>a:hover span.toctree-expand{color:gray}.wy-menu-vertical li.on a span.toctree-expand,.wy-menu-vertical li.current>a span.toctree-expand{display:block;font-size:.8em;line-height:1.6em;color:#333}.wy-menu-vertical li.toctree-l1.current>a{border-bottom:solid 1px #c9c9c9;border-top:solid 1px #c9c9c9}.wy-menu-vertical li.toctree-l2 a,.wy-menu-vertical li.toctree-l3 a,.wy-menu-vertical li.toctree-l4 a{color:#404040}.wy-menu-vertical li.toctree-l1.current li.toctree-l2>ul,.wy-menu-vertical li.toctree-l2.current li.toctree-l3>ul{display:none}.wy-menu-vertical li.toctree-l1.current li.toctree-l2.current>ul,.wy-menu-vertical li.toctree-l2.current li.toctree-l3.current>ul{display:block}.wy-menu-vertical li.toctree-l2.current>a{background:#c9c9c9;padding:.4045em 2.427em}.wy-menu-vertical li.toctree-l2.current li.toctree-l3>a{display:block;background:#c9c9c9;padding:.4045em 4.045em}.wy-menu-vertical li.toctree-l2 a:hover span.toctree-expand{color:gray}.wy-menu-vertical li.toctree-l2 span.toctree-expand{color:#a3a3a3}.wy-menu-vertical li.toctree-l3{font-size:.9em}.wy-menu-vertical li.toctree-l3.current>a{background:#bdbdbd;padding:.4045em 4.045em}.wy-menu-vertical li.toctree-l3.current li.toctree-l4>a{display:block;background:#bdbdbd;padding:.4045em 5.663em}.wy-menu-vertical li.toctree-l3 a:hover span.toctree-expand{color:gray}.wy-menu-vertical li.toctree-l3 span.toctree-expand{color:#969696}.wy-menu-vertical li.toctree-l4{font-size:.9em}.wy-menu-vertical li.current ul{display:block}.wy-menu-vertical li ul{margin-bottom:0;display:none}.wy-menu-vertical li ul li a{margin-bottom:0;color:#d9d9d9;font-weight:normal}.wy-menu-vertical a{display:inline-block;line-height:18px;padding:.4045em 1.618em;display:block;position:relative;font-size:90%;color:#d9d9d9}.wy-menu-vertical a:hover{background-color:#4e4a4a;cursor:pointer}.wy-menu-vertical a:hover span.toctree-expand{color:#d9d9d9}.wy-menu-vertical a:active{background-color:#2980B9;cursor:pointer;color:#fff}.wy-menu-vertical a:active span.toctree-expand{color:#fff}.wy-side-nav-search{display:block;width:300px;padding:.809em;margin-bottom:.809em;z-index:200;background-color:#2980B9;text-align:center;padding:.809em;display:block;color:#fcfcfc;margin-bottom:.809em}.wy-side-nav-search input[type=text]{width:100%;border-radius:50px;padding:6px 12px;border-color:#2472a4}.wy-side-nav-search img{display:block;margin:auto auto .809em auto;height:45px;width:45px;background-color:#2980B9;padding:5px;border-radius:100%}.wy-side-nav-search>a,.wy-side-nav-search .wy-dropdown>a{color:#fcfcfc;font-size:100%;font-weight:bold;display:inline-block;padding:4px 6px;margin-bottom:.809em}.wy-side-nav-search>a:hover,.wy-side-nav-search .wy-dropdown>a:hover{background:rgba(255,255,255,0.1)}.wy-side-nav-search>a img.logo,.wy-side-nav-search .wy-dropdown>a img.logo{display:block;margin:0 auto;height:auto;width:auto;border-radius:0;max-width:100%;background:transparent}.wy-side-nav-search>a.icon img.logo,.wy-side-nav-search .wy-dropdown>a.icon img.logo{margin-top:.85em}.wy-side-nav-search>div.version{margin-top:-.4045em;margin-bottom:.809em;font-weight:normal;color:rgba(255,255,255,0.3)}.wy-nav .wy-menu-vertical header{color:#2980B9}.wy-nav .wy-menu-vertical a{color:#b3b3b3}.wy-nav .wy-menu-vertical a:hover{background-color:#2980B9;color:#fff}[data-menu-wrap]{-webkit-transition:all .2s ease-in;-moz-transition:all .2s ease-in;transition:all .2s ease-in;position:absolute;opacity:1;width:100%;opacity:0}[data-menu-wrap].move-center{left:0;right:auto;opacity:1}[data-menu-wrap].move-left{right:auto;left:-100%;opacity:0}[data-menu-wrap].move-right{right:-100%;left:auto;opacity:0}.wy-body-for-nav{background:#fcfcfc}.wy-grid-for-nav{position:absolute;width:100%;height:100%}.wy-nav-side{position:fixed;top:0;bottom:0;left:0;padding-bottom:2em;width:300px;overflow-x:hidden;overflow-y:hidden;min-height:100%;color:#9b9b9b;background:#343131;z-index:200}.wy-side-scroll{width:320px;position:relative;overflow-x:hidden;overflow-y:scroll;height:100%}.wy-nav-top{display:none;background:#2980B9;color:#fff;padding:.4045em .809em;position:relative;line-height:50px;text-align:center;font-size:100%;*zoom:1}.wy-nav-top:before,.wy-nav-top:after{display:table;content:""}.wy-nav-top:after{clear:both}.wy-nav-top a{color:#fff;font-weight:bold}.wy-nav-top img{margin-right:12px;height:45px;width:45px;background-color:#2980B9;padding:5px;border-radius:100%}.wy-nav-top i{font-size:30px;float:left;cursor:pointer;padding-top:inherit}.wy-nav-content-wrap{margin-left:300px;background:#fcfcfc;min-height:100%}.wy-nav-content{padding:1.618em 3.236em;height:100%;max-width:800px;margin:auto}.wy-body-mask{position:fixed;width:100%;height:100%;background:rgba(0,0,0,0.2);display:none;z-index:499}.wy-body-mask.on{display:block}footer{color:gray}footer p{margin-bottom:12px}footer span.commit code,footer span.commit .rst-content tt,.rst-content footer span.commit tt{padding:0px;font-family:SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace;font-size:1em;background:none;border:none;color:gray}.rst-footer-buttons{*zoom:1}.rst-footer-buttons:before,.rst-footer-buttons:after{width:100%}.rst-footer-buttons:before,.rst-footer-buttons:after{display:table;content:""}.rst-footer-buttons:after{clear:both}.rst-breadcrumbs-buttons{margin-top:12px;*zoom:1}.rst-breadcrumbs-buttons:before,.rst-breadcrumbs-buttons:after{display:table;content:""}.rst-breadcrumbs-buttons:after{clear:both}#search-results .search li{margin-bottom:24px;border-bottom:solid 1px #e1e4e5;padding-bottom:24px}#search-results .search li:first-child{border-top:solid 1px #e1e4e5;padding-top:24px}#search-results .search li a{font-size:120%;margin-bottom:12px;display:inline-block}#search-results .context{color:gray;font-size:90%}@media screen and (max-width: 768px){.wy-body-for-nav{background:#fcfcfc}.wy-nav-top{display:block}.wy-nav-side{left:-300px}.wy-nav-side.shift{width:85%;left:0}.wy-side-scroll{width:auto}.wy-side-nav-search{width:auto}.wy-menu.wy-menu-vertical{width:auto}.wy-nav-content-wrap{margin-left:0}.wy-nav-content-wrap .wy-nav-content{padding:1.618em}.wy-nav-content-wrap.shift{position:fixed;min-width:100%;left:85%;top:0;height:100%;overflow:hidden}}@media screen and (min-width: 1100px){.wy-nav-content-wrap{background:rgba(0,0,0,0.05)}.wy-nav-content{margin:0;background:#fcfcfc}}@media print{.rst-versions,footer,.wy-nav-side{display:none}.wy-nav-content-wrap{margin-left:0}}.rst-versions{position:fixed;bottom:0;left:0;width:300px;color:#fcfcfc;background:#1f1d1d;font-family:"Lato","proxima-nova","Helvetica Neue",Arial,sans-serif;z-index:400}.rst-versions a{color:#2980B9;text-decoration:none}.rst-versions .rst-badge-small{display:none}.rst-versions .rst-current-version{padding:12px;background-color:#272525;display:block;text-align:right;font-size:90%;cursor:pointer;color:#27AE60;*zoom:1}.rst-versions .rst-current-version:before,.rst-versions .rst-current-version:after{display:table;content:""}.rst-versions .rst-current-version:after{clear:both}.rst-versions .rst-current-version .fa,.rst-versions .rst-current-version .wy-menu-vertical li span.toctree-expand,.wy-menu-vertical li .rst-versions .rst-current-version span.toctree-expand,.rst-versions .rst-current-version .rst-content .admonition-title,.rst-content .rst-versions .rst-current-version .admonition-title,.rst-versions .rst-current-version .rst-content h1 .headerlink,.rst-content h1 .rst-versions .rst-current-version .headerlink,.rst-versions .rst-current-version .rst-content h2 .headerlink,.rst-content h2 .rst-versions .rst-current-version .headerlink,.rst-versions .rst-current-version .rst-content h3 .headerlink,.rst-content h3 .rst-versions .rst-current-version .headerlink,.rst-versions .rst-current-version .rst-content h4 .headerlink,.rst-content h4 .rst-versions .rst-current-version .headerlink,.rst-versions .rst-current-version .rst-content h5 .headerlink,.rst-content h5 .rst-versions .rst-current-version .headerlink,.rst-versions .rst-current-version .rst-content h6 .headerlink,.rst-content h6 .rst-versions .rst-current-version .headerlink,.rst-versions .rst-current-version .rst-content dl dt .headerlink,.rst-content dl dt .rst-versions .rst-current-version .headerlink,.rst-versions .rst-current-version .rst-content p.caption .headerlink,.rst-content p.caption .rst-versions .rst-current-version .headerlink,.rst-versions .rst-current-version .rst-content table>caption .headerlink,.rst-content table>caption .rst-versions .rst-current-version .headerlink,.rst-versions .rst-current-version .rst-content tt.download span:first-child,.rst-content tt.download .rst-versions .rst-current-version span:first-child,.rst-versions .rst-current-version .rst-content code.download span:first-child,.rst-content code.download .rst-versions .rst-current-version span:first-child,.rst-versions .rst-current-version .icon{color:#fcfcfc}.rst-versions .rst-current-version .fa-book,.rst-versions .rst-current-version .icon-book{float:left}.rst-versions .rst-current-version .icon-book{float:left}.rst-versions .rst-current-version.rst-out-of-date{background-color:#E74C3C;color:#fff}.rst-versions .rst-current-version.rst-active-old-version{background-color:#F1C40F;color:#000}.rst-versions.shift-up{height:auto;max-height:100%}.rst-versions.shift-up .rst-other-versions{display:block}.rst-versions .rst-other-versions{font-size:90%;padding:12px;color:gray;display:none}.rst-versions .rst-other-versions hr{display:block;height:1px;border:0;margin:20px 0;padding:0;border-top:solid 1px #413d3d}.rst-versions .rst-other-versions dd{display:inline-block;margin:0}.rst-versions .rst-other-versions dd a{display:inline-block;padding:6px;color:#fcfcfc}.rst-versions.rst-badge{width:auto;bottom:20px;right:20px;left:auto;border:none;max-width:300px}.rst-versions.rst-badge .icon-book{float:none}.rst-versions.rst-badge .fa-book,.rst-versions.rst-badge .icon-book{float:none}.rst-versions.rst-badge.shift-up .rst-current-version{text-align:right}.rst-versions.rst-badge.shift-up .rst-current-version .fa-book,.rst-versions.rst-badge.shift-up .rst-current-version .icon-book{float:left}.rst-versions.rst-badge.shift-up .rst-current-version .icon-book{float:left}.rst-versions.rst-badge .rst-current-version{width:auto;height:30px;line-height:30px;padding:0 6px;display:block;text-align:center}@media screen and (max-width: 768px){.rst-versions{width:85%;display:none}.rst-versions.shift{display:block}}.rst-content img{max-width:100%;height:auto}.rst-content div.figure{margin-bottom:24px}.rst-content div.figure p.caption{font-style:italic}.rst-content div.figure p:last-child.caption{margin-bottom:0px}.rst-content div.figure.align-center{text-align:center}.rst-content .section>img,.rst-content .section>a>img{margin-bottom:24px}.rst-content abbr[title]{text-decoration:none}.rst-content.style-external-links a.reference.external:after{font-family:FontAwesome;content:"";color:#b3b3b3;vertical-align:super;font-size:60%;margin:0 .2em}.rst-content blockquote{margin-left:24px;line-height:24px;margin-bottom:24px}.rst-content pre.literal-block{white-space:pre;margin:0;padding:12px 12px;font-family:SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace;display:block;overflow:auto}.rst-content pre.literal-block,.rst-content div[class^='highlight']{border:1px solid #e1e4e5;overflow-x:auto;margin:1px 0 24px 0}.rst-content pre.literal-block div[class^='highlight'],.rst-content div[class^='highlight'] div[class^='highlight']{padding:0px;border:none;margin:0}.rst-content div[class^='highlight'] td.code{width:100%}.rst-content .linenodiv pre{border-right:solid 1px #e6e9ea;margin:0;padding:12px 12px;font-family:SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace;user-select:none;pointer-events:none}.rst-content div[class^='highlight'] pre{white-space:pre;margin:0;padding:12px 12px;display:block;overflow:auto}.rst-content div[class^='highlight'] pre .hll{display:block;margin:0 -12px;padding:0 12px}.rst-content pre.literal-block,.rst-content div[class^='highlight'] pre,.rst-content .linenodiv pre{font-family:SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace;font-size:12px;line-height:1.4}@media print{.rst-content .codeblock,.rst-content div[class^='highlight'],.rst-content div[class^='highlight'] pre{white-space:pre-wrap}}.rst-content .note .last,.rst-content .attention .last,.rst-content .caution .last,.rst-content .danger .last,.rst-content .error .last,.rst-content .hint .last,.rst-content .important .last,.rst-content .tip .last,.rst-content .warning .last,.rst-content .seealso .last,.rst-content .admonition-todo .last,.rst-content .admonition .last{margin-bottom:0}.rst-content .admonition-title:before{margin-right:4px}.rst-content .admonition table{border-color:rgba(0,0,0,0.1)}.rst-content .admonition table td,.rst-content .admonition table th{background:transparent !important;border-color:rgba(0,0,0,0.1) !important}.rst-content .section ol.loweralpha,.rst-content .section ol.loweralpha li{list-style:lower-alpha}.rst-content .section ol.upperalpha,.rst-content .section ol.upperalpha li{list-style:upper-alpha}.rst-content .section ol p,.rst-content .section ul p{margin-bottom:12px}.rst-content .section ol p:last-child,.rst-content .section ul p:last-child{margin-bottom:24px}.rst-content .line-block{margin-left:0px;margin-bottom:24px;line-height:24px}.rst-content .line-block .line-block{margin-left:24px;margin-bottom:0px}.rst-content .topic-title{font-weight:bold;margin-bottom:12px}.rst-content .toc-backref{color:#404040}.rst-content .align-right{float:right;margin:0px 0px 24px 24px}.rst-content .align-left{float:left;margin:0px 24px 24px 0px}.rst-content .align-center{margin:auto}.rst-content .align-center:not(table){display:block}.rst-content h1 .headerlink,.rst-content h2 .headerlink,.rst-content .toctree-wrapper p.caption .headerlink,.rst-content h3 .headerlink,.rst-content h4 .headerlink,.rst-content h5 .headerlink,.rst-content h6 .headerlink,.rst-content dl dt .headerlink,.rst-content p.caption .headerlink,.rst-content table>caption .headerlink{visibility:hidden;font-size:14px}.rst-content h1 .headerlink:after,.rst-content h2 .headerlink:after,.rst-content .toctree-wrapper p.caption .headerlink:after,.rst-content h3 .headerlink:after,.rst-content h4 .headerlink:after,.rst-content h5 .headerlink:after,.rst-content h6 .headerlink:after,.rst-content dl dt .headerlink:after,.rst-content p.caption .headerlink:after,.rst-content table>caption .headerlink:after{content:"";font-family:FontAwesome}.rst-content h1:hover .headerlink:after,.rst-content h2:hover .headerlink:after,.rst-content .toctree-wrapper p.caption:hover .headerlink:after,.rst-content h3:hover .headerlink:after,.rst-content h4:hover .headerlink:after,.rst-content h5:hover .headerlink:after,.rst-content h6:hover .headerlink:after,.rst-content dl dt:hover .headerlink:after,.rst-content p.caption:hover .headerlink:after,.rst-content table>caption:hover .headerlink:after{visibility:visible}.rst-content table>caption .headerlink:after{font-size:12px}.rst-content .centered{text-align:center}.rst-content .sidebar{float:right;width:40%;display:block;margin:0 0 24px 24px;padding:24px;background:#f3f6f6;border:solid 1px #e1e4e5}.rst-content .sidebar p,.rst-content .sidebar ul,.rst-content .sidebar dl{font-size:90%}.rst-content .sidebar .last{margin-bottom:0}.rst-content .sidebar .sidebar-title{display:block;font-family:"Roboto Slab","ff-tisa-web-pro","Georgia",Arial,sans-serif;font-weight:bold;background:#e1e4e5;padding:6px 12px;margin:-24px;margin-bottom:24px;font-size:100%}.rst-content .highlighted{background:#F1C40F;display:inline-block;font-weight:bold;padding:0 6px}.rst-content .footnote-reference,.rst-content .citation-reference{vertical-align:baseline;position:relative;top:-0.4em;line-height:0;font-size:90%}.rst-content table.docutils.citation,.rst-content table.docutils.footnote{background:none;border:none;color:gray}.rst-content table.docutils.citation td,.rst-content table.docutils.citation tr,.rst-content table.docutils.footnote td,.rst-content table.docutils.footnote tr{border:none;background-color:transparent !important;white-space:normal}.rst-content table.docutils.citation td.label,.rst-content table.docutils.footnote td.label{padding-left:0;padding-right:0;vertical-align:top}.rst-content table.docutils.citation tt,.rst-content table.docutils.citation code,.rst-content table.docutils.footnote tt,.rst-content table.docutils.footnote code{color:#555}.rst-content .wy-table-responsive.citation,.rst-content .wy-table-responsive.footnote{margin-bottom:0}.rst-content .wy-table-responsive.citation+:not(.citation),.rst-content .wy-table-responsive.footnote+:not(.footnote){margin-top:24px}.rst-content .wy-table-responsive.citation:last-child,.rst-content .wy-table-responsive.footnote:last-child{margin-bottom:24px}.rst-content table.docutils th{border-color:#e1e4e5}.rst-content table.docutils td .last,.rst-content table.docutils td .last :last-child{margin-bottom:0}.rst-content table.field-list{border:none}.rst-content table.field-list td{border:none}.rst-content table.field-list td>strong{display:inline-block}.rst-content table.field-list .field-name{padding-right:10px;text-align:left;white-space:nowrap}.rst-content table.field-list .field-body{text-align:left}.rst-content tt,.rst-content tt,.rst-content code{color:#000;font-family:SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace;padding:2px 5px}.rst-content tt big,.rst-content tt em,.rst-content tt big,.rst-content code big,.rst-content tt em,.rst-content code em{font-size:100% !important;line-height:normal}.rst-content tt.literal,.rst-content tt.literal,.rst-content code.literal{color:#E74C3C}.rst-content tt.xref,a .rst-content tt,.rst-content tt.xref,.rst-content code.xref,a .rst-content tt,a .rst-content code{font-weight:bold;color:#404040}.rst-content pre,.rst-content kbd,.rst-content samp{font-family:SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",Courier,monospace}.rst-content a tt,.rst-content a tt,.rst-content a code{color:#2980B9}.rst-content dl{margin-bottom:24px}.rst-content dl dt{font-weight:bold;margin-bottom:12px}.rst-content dl p,.rst-content dl table,.rst-content dl ul,.rst-content dl ol{margin-bottom:12px !important}.rst-content dl dd{margin:0 0 12px 24px;line-height:24px}.rst-content dl:not(.docutils){margin-bottom:24px}.rst-content dl:not(.docutils) dt{display:table;margin:6px 0;font-size:90%;line-height:normal;background:#e7f2fa;color:#2980B9;border-top:solid 3px #6ab0de;padding:6px;position:relative}.rst-content dl:not(.docutils) dt:before{color:#6ab0de}.rst-content dl:not(.docutils) dt .headerlink{color:#404040;font-size:100% !important}.rst-content dl:not(.docutils) dl dt{margin-bottom:6px;border:none;border-left:solid 3px #ccc;background:#f0f0f0;color:#555}.rst-content dl:not(.docutils) dl dt .headerlink{color:#404040;font-size:100% !important}.rst-content dl:not(.docutils) dt:first-child{margin-top:0}.rst-content dl:not(.docutils) tt,.rst-content dl:not(.docutils) tt,.rst-content dl:not(.docutils) code{font-weight:bold}.rst-content dl:not(.docutils) tt.descname,.rst-content dl:not(.docutils) tt.descclassname,.rst-content dl:not(.docutils) tt.descname,.rst-content dl:not(.docutils) code.descname,.rst-content dl:not(.docutils) tt.descclassname,.rst-content dl:not(.docutils) code.descclassname{background-color:transparent;border:none;padding:0;font-size:100% !important}.rst-content dl:not(.docutils) tt.descname,.rst-content dl:not(.docutils) tt.descname,.rst-content dl:not(.docutils) code.descname{font-weight:bold}.rst-content dl:not(.docutils) .optional{display:inline-block;padding:0 4px;color:#000;font-weight:bold}.rst-content dl:not(.docutils) .property{display:inline-block;padding-right:8px}.rst-content .viewcode-link,.rst-content .viewcode-back{display:inline-block;color:#27AE60;font-size:80%;padding-left:24px}.rst-content .viewcode-back{display:block;float:right}.rst-content p.rubric{margin-bottom:12px;font-weight:bold}.rst-content tt.download,.rst-content code.download{background:inherit;padding:inherit;font-weight:normal;font-family:inherit;font-size:inherit;color:inherit;border:inherit;white-space:inherit}.rst-content tt.download span:first-child,.rst-content code.download span:first-child{-webkit-font-smoothing:subpixel-antialiased}.rst-content tt.download span:first-child:before,.rst-content code.download span:first-child:before{margin-right:4px}.rst-content .guilabel{border:1px solid #7fbbe3;background:#e7f2fa;font-size:80%;font-weight:700;border-radius:4px;padding:2.4px 6px;margin:auto 2px}.rst-content .versionmodified{font-style:italic}@media screen and (max-width: 480px){.rst-content .sidebar{width:100%}}span[id*='MathJax-Span']{color:#404040}.math{text-align:center}@font-face{font-family:"Lato";src:url("../fonts/Lato/lato-regular.eot");src:url("../fonts/Lato/lato-regular.eot?#iefix") format("embedded-opentype"),url("../fonts/Lato/lato-regular.woff2") format("woff2"),url("../fonts/Lato/lato-regular.woff") format("woff"),url("../fonts/Lato/lato-regular.ttf") format("truetype");font-weight:400;font-style:normal}@font-face{font-family:"Lato";src:url("../fonts/Lato/lato-bold.eot");src:url("../fonts/Lato/lato-bold.eot?#iefix") format("embedded-opentype"),url("../fonts/Lato/lato-bold.woff2") format("woff2"),url("../fonts/Lato/lato-bold.woff") format("woff"),url("../fonts/Lato/lato-bold.ttf") format("truetype");font-weight:700;font-style:normal}@font-face{font-family:"Lato";src:url("../fonts/Lato/lato-bolditalic.eot");src:url("../fonts/Lato/lato-bolditalic.eot?#iefix") format("embedded-opentype"),url("../fonts/Lato/lato-bolditalic.woff2") format("woff2"),url("../fonts/Lato/lato-bolditalic.woff") format("woff"),url("../fonts/Lato/lato-bolditalic.ttf") format("truetype");font-weight:700;font-style:italic}@font-face{font-family:"Lato";src:url("../fonts/Lato/lato-italic.eot");src:url("../fonts/Lato/lato-italic.eot?#iefix") format("embedded-opentype"),url("../fonts/Lato/lato-italic.woff2") format("woff2"),url("../fonts/Lato/lato-italic.woff") format("woff"),url("../fonts/Lato/lato-italic.ttf") format("truetype");font-weight:400;font-style:italic}@font-face{font-family:"Roboto Slab";font-style:normal;font-weight:400;src:url("../fonts/RobotoSlab/roboto-slab.eot");src:url("../fonts/RobotoSlab/roboto-slab-v7-regular.eot?#iefix") format("embedded-opentype"),url("../fonts/RobotoSlab/roboto-slab-v7-regular.woff2") format("woff2"),url("../fonts/RobotoSlab/roboto-slab-v7-regular.woff") format("woff"),url("../fonts/RobotoSlab/roboto-slab-v7-regular.ttf") format("truetype")}@font-face{font-family:"Roboto Slab";font-style:normal;font-weight:700;src:url("../fonts/RobotoSlab/roboto-slab-v7-bold.eot");src:url("../fonts/RobotoSlab/roboto-slab-v7-bold.eot?#iefix") format("embedded-opentype"),url("../fonts/RobotoSlab/roboto-slab-v7-bold.woff2") format("woff2"),url("../fonts/RobotoSlab/roboto-slab-v7-bold.woff") format("woff"),url("../fonts/RobotoSlab/roboto-slab-v7-bold.ttf") format("truetype")} diff --git a/docs/_build/html/_static/doctools.js b/docs/_build/html/_static/doctools.js new file mode 100644 index 000000000..d8928926b --- /dev/null +++ b/docs/_build/html/_static/doctools.js @@ -0,0 +1,313 @@ +/* + * doctools.js + * ~~~~~~~~~~~ + * + * Sphinx JavaScript utilities for all documentation. + * + * :copyright: Copyright 2007-2018 by the Sphinx team, see AUTHORS. + * :license: BSD, see LICENSE for details. + * + */ + +/** + * select a different prefix for underscore + */ +$u = _.noConflict(); + +/** + * make the code below compatible with browsers without + * an installed firebug like debugger +if (!window.console || !console.firebug) { + var names = ["log", "debug", "info", "warn", "error", "assert", "dir", + "dirxml", "group", "groupEnd", "time", "timeEnd", "count", "trace", + "profile", "profileEnd"]; + window.console = {}; + for (var i = 0; i < names.length; ++i) + window.console[names[i]] = function() {}; +} + */ + +/** + * small helper function to urldecode strings + */ +jQuery.urldecode = function(x) { + return decodeURIComponent(x).replace(/\+/g, ' '); +}; + +/** + * small helper function to urlencode strings + */ +jQuery.urlencode = encodeURIComponent; + +/** + * This function returns the parsed url parameters of the + * current request. Multiple values per key are supported, + * it will always return arrays of strings for the value parts. + */ +jQuery.getQueryParameters = function(s) { + if (typeof s === 'undefined') + s = document.location.search; + var parts = s.substr(s.indexOf('?') + 1).split('&'); + var result = {}; + for (var i = 0; i < parts.length; i++) { + var tmp = parts[i].split('=', 2); + var key = jQuery.urldecode(tmp[0]); + var value = jQuery.urldecode(tmp[1]); + if (key in result) + result[key].push(value); + else + result[key] = [value]; + } + return result; +}; + +/** + * highlight a given string on a jquery object by wrapping it in + * span elements with the given class name. + */ +jQuery.fn.highlightText = function(text, className) { + function highlight(node, addItems) { + if (node.nodeType === 3) { + var val = node.nodeValue; + var pos = val.toLowerCase().indexOf(text); + if (pos >= 0 && + !jQuery(node.parentNode).hasClass(className) && + !jQuery(node.parentNode).hasClass("nohighlight")) { + var span; + var isInSVG = jQuery(node).closest("body, svg, foreignObject").is("svg"); + if (isInSVG) { + span = document.createElementNS("http://www.w3.org/2000/svg", "tspan"); + } else { + span = document.createElement("span"); + span.className = className; + } + span.appendChild(document.createTextNode(val.substr(pos, text.length))); + node.parentNode.insertBefore(span, node.parentNode.insertBefore( + document.createTextNode(val.substr(pos + text.length)), + node.nextSibling)); + node.nodeValue = val.substr(0, pos); + if (isInSVG) { + var bbox = span.getBBox(); + var rect = document.createElementNS("http://www.w3.org/2000/svg", "rect"); + rect.x.baseVal.value = bbox.x; + rect.y.baseVal.value = bbox.y; + rect.width.baseVal.value = bbox.width; + rect.height.baseVal.value = bbox.height; + rect.setAttribute('class', className); + var parentOfText = node.parentNode.parentNode; + addItems.push({ + "parent": node.parentNode, + "target": rect}); + } + } + } + else if (!jQuery(node).is("button, select, textarea")) { + jQuery.each(node.childNodes, function() { + highlight(this, addItems); + }); + } + } + var addItems = []; + var result = this.each(function() { + highlight(this, addItems); + }); + for (var i = 0; i < addItems.length; ++i) { + jQuery(addItems[i].parent).before(addItems[i].target); + } + return result; +}; + +/* + * backward compatibility for jQuery.browser + * This will be supported until firefox bug is fixed. + */ +if (!jQuery.browser) { + jQuery.uaMatch = function(ua) { + ua = ua.toLowerCase(); + + var match = /(chrome)[ \/]([\w.]+)/.exec(ua) || + /(webkit)[ \/]([\w.]+)/.exec(ua) || + /(opera)(?:.*version|)[ \/]([\w.]+)/.exec(ua) || + /(msie) ([\w.]+)/.exec(ua) || + ua.indexOf("compatible") < 0 && /(mozilla)(?:.*? rv:([\w.]+)|)/.exec(ua) || + []; + + return { + browser: match[ 1 ] || "", + version: match[ 2 ] || "0" + }; + }; + jQuery.browser = {}; + jQuery.browser[jQuery.uaMatch(navigator.userAgent).browser] = true; +} + +/** + * Small JavaScript module for the documentation. + */ +var Documentation = { + + init : function() { + this.fixFirefoxAnchorBug(); + this.highlightSearchWords(); + this.initIndexTable(); + + }, + + /** + * i18n support + */ + TRANSLATIONS : {}, + PLURAL_EXPR : function(n) { return n === 1 ? 0 : 1; }, + LOCALE : 'unknown', + + // gettext and ngettext don't access this so that the functions + // can safely bound to a different name (_ = Documentation.gettext) + gettext : function(string) { + var translated = Documentation.TRANSLATIONS[string]; + if (typeof translated === 'undefined') + return string; + return (typeof translated === 'string') ? translated : translated[0]; + }, + + ngettext : function(singular, plural, n) { + var translated = Documentation.TRANSLATIONS[singular]; + if (typeof translated === 'undefined') + return (n == 1) ? singular : plural; + return translated[Documentation.PLURALEXPR(n)]; + }, + + addTranslations : function(catalog) { + for (var key in catalog.messages) + this.TRANSLATIONS[key] = catalog.messages[key]; + this.PLURAL_EXPR = new Function('n', 'return +(' + catalog.plural_expr + ')'); + this.LOCALE = catalog.locale; + }, + + /** + * add context elements like header anchor links + */ + addContextElements : function() { + $('div[id] > :header:first').each(function() { + $('\u00B6'). + attr('href', '#' + this.id). + attr('title', _('Permalink to this headline')). + appendTo(this); + }); + $('dt[id]').each(function() { + $('\u00B6'). + attr('href', '#' + this.id). + attr('title', _('Permalink to this definition')). + appendTo(this); + }); + }, + + /** + * workaround a firefox stupidity + * see: https://bugzilla.mozilla.org/show_bug.cgi?id=645075 + */ + fixFirefoxAnchorBug : function() { + if (document.location.hash && $.browser.mozilla) + window.setTimeout(function() { + document.location.href += ''; + }, 10); + }, + + /** + * highlight the search words provided in the url in the text + */ + highlightSearchWords : function() { + var params = $.getQueryParameters(); + var terms = (params.highlight) ? params.highlight[0].split(/\s+/) : []; + if (terms.length) { + var body = $('div.body'); + if (!body.length) { + body = $('body'); + } + window.setTimeout(function() { + $.each(terms, function() { + body.highlightText(this.toLowerCase(), 'highlighted'); + }); + }, 10); + $('') + .appendTo($('#searchbox')); + } + }, + + /** + * init the domain index toggle buttons + */ + initIndexTable : function() { + var togglers = $('img.toggler').click(function() { + var src = $(this).attr('src'); + var idnum = $(this).attr('id').substr(7); + $('tr.cg-' + idnum).toggle(); + if (src.substr(-9) === 'minus.png') + $(this).attr('src', src.substr(0, src.length-9) + 'plus.png'); + else + $(this).attr('src', src.substr(0, src.length-8) + 'minus.png'); + }).css('display', ''); + if (DOCUMENTATION_OPTIONS.COLLAPSE_INDEX) { + togglers.click(); + } + }, + + /** + * helper function to hide the search marks again + */ + hideSearchWords : function() { + $('#searchbox .highlight-link').fadeOut(300); + $('span.highlighted').removeClass('highlighted'); + }, + + /** + * make the url absolute + */ + makeURL : function(relativeURL) { + return DOCUMENTATION_OPTIONS.URL_ROOT + '/' + relativeURL; + }, + + /** + * get the current relative url + */ + getCurrentURL : function() { + var path = document.location.pathname; + var parts = path.split(/\//); + $.each(DOCUMENTATION_OPTIONS.URL_ROOT.split(/\//), function() { + if (this === '..') + parts.pop(); + }); + var url = parts.join('/'); + return path.substring(url.lastIndexOf('/') + 1, path.length - 1); + }, + + initOnKeyListeners: function() { + $(document).keyup(function(event) { + var activeElementType = document.activeElement.tagName; + // don't navigate when in search box or textarea + if (activeElementType !== 'TEXTAREA' && activeElementType !== 'INPUT' && activeElementType !== 'SELECT') { + switch (event.keyCode) { + case 37: // left + var prevHref = $('link[rel="prev"]').prop('href'); + if (prevHref) { + window.location.href = prevHref; + return false; + } + case 39: // right + var nextHref = $('link[rel="next"]').prop('href'); + if (nextHref) { + window.location.href = nextHref; + return false; + } + } + } + }); + } +}; + +// quick alias for translations +_ = Documentation.gettext; + +$(document).ready(function() { + Documentation.init(); +}); \ No newline at end of file diff --git a/docs/_build/html/_static/documentation_options.js b/docs/_build/html/_static/documentation_options.js new file mode 100644 index 000000000..4e68537a7 --- /dev/null +++ b/docs/_build/html/_static/documentation_options.js @@ -0,0 +1,9 @@ +var DOCUMENTATION_OPTIONS = { + URL_ROOT: document.getElementById("documentation_options").getAttribute('data-url_root'), + VERSION: '1.7', + LANGUAGE: 'None', + COLLAPSE_INDEX: false, + FILE_SUFFIX: '.html', + HAS_SOURCE: true, + SOURCELINK_SUFFIX: '.txt' +}; \ No newline at end of file diff --git a/docs/_build/html/_static/down-pressed.png b/docs/_build/html/_static/down-pressed.png new file mode 100644 index 000000000..5756c8cad Binary files /dev/null and b/docs/_build/html/_static/down-pressed.png differ diff --git a/docs/_build/html/_static/down.png b/docs/_build/html/_static/down.png new file mode 100644 index 000000000..1b3bdad2c Binary files /dev/null and b/docs/_build/html/_static/down.png differ diff --git a/docs/_build/html/_static/file.png b/docs/_build/html/_static/file.png new file mode 100644 index 000000000..a858a410e Binary files /dev/null and b/docs/_build/html/_static/file.png differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-bold.eot b/docs/_build/html/_static/fonts/Lato/lato-bold.eot new file mode 100644 index 000000000..3361183a4 Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-bold.eot differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-bold.ttf b/docs/_build/html/_static/fonts/Lato/lato-bold.ttf new file mode 100644 index 000000000..29f691d5e Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-bold.ttf differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-bold.woff b/docs/_build/html/_static/fonts/Lato/lato-bold.woff new file mode 100644 index 000000000..c6dff51f0 Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-bold.woff differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-bold.woff2 b/docs/_build/html/_static/fonts/Lato/lato-bold.woff2 new file mode 100644 index 000000000..bb195043c Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-bold.woff2 differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-bolditalic.eot b/docs/_build/html/_static/fonts/Lato/lato-bolditalic.eot new file mode 100644 index 000000000..3d4154936 Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-bolditalic.eot differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-bolditalic.ttf b/docs/_build/html/_static/fonts/Lato/lato-bolditalic.ttf new file mode 100644 index 000000000..f402040b3 Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-bolditalic.ttf differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-bolditalic.woff b/docs/_build/html/_static/fonts/Lato/lato-bolditalic.woff new file mode 100644 index 000000000..88ad05b9f Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-bolditalic.woff differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-bolditalic.woff2 b/docs/_build/html/_static/fonts/Lato/lato-bolditalic.woff2 new file mode 100644 index 000000000..c4e3d804b Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-bolditalic.woff2 differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-italic.eot b/docs/_build/html/_static/fonts/Lato/lato-italic.eot new file mode 100644 index 000000000..3f826421a Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-italic.eot differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-italic.ttf b/docs/_build/html/_static/fonts/Lato/lato-italic.ttf new file mode 100644 index 000000000..b4bfc9b24 Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-italic.ttf differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-italic.woff b/docs/_build/html/_static/fonts/Lato/lato-italic.woff new file mode 100644 index 000000000..76114bc03 Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-italic.woff differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-italic.woff2 b/docs/_build/html/_static/fonts/Lato/lato-italic.woff2 new file mode 100644 index 000000000..3404f37e2 Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-italic.woff2 differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-regular.eot b/docs/_build/html/_static/fonts/Lato/lato-regular.eot new file mode 100644 index 000000000..11e3f2a5f Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-regular.eot differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-regular.ttf b/docs/_build/html/_static/fonts/Lato/lato-regular.ttf new file mode 100644 index 000000000..74decd9eb Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-regular.ttf differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-regular.woff b/docs/_build/html/_static/fonts/Lato/lato-regular.woff new file mode 100644 index 000000000..ae1307ff5 Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-regular.woff differ diff --git a/docs/_build/html/_static/fonts/Lato/lato-regular.woff2 b/docs/_build/html/_static/fonts/Lato/lato-regular.woff2 new file mode 100644 index 000000000..3bf984332 Binary files /dev/null and b/docs/_build/html/_static/fonts/Lato/lato-regular.woff2 differ diff --git a/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.eot b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.eot new file mode 100644 index 000000000..79dc8efed Binary files /dev/null and b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.eot differ diff --git a/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.ttf b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.ttf new file mode 100644 index 000000000..df5d1df27 Binary files /dev/null and b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.ttf differ diff --git a/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.woff b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.woff new file mode 100644 index 000000000..6cb600001 Binary files /dev/null and b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.woff differ diff --git a/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.woff2 b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.woff2 new file mode 100644 index 000000000..7059e2314 Binary files /dev/null and b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-bold.woff2 differ diff --git a/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.eot b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.eot new file mode 100644 index 000000000..2f7ca78a1 Binary files /dev/null and b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.eot differ diff --git a/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.ttf b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.ttf new file mode 100644 index 000000000..eb52a7907 Binary files /dev/null and b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.ttf differ diff --git a/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.woff b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.woff new file mode 100644 index 000000000..f815f63f9 Binary files /dev/null and b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.woff differ diff --git a/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.woff2 b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.woff2 new file mode 100644 index 000000000..f2c76e5bd Binary files /dev/null and b/docs/_build/html/_static/fonts/RobotoSlab/roboto-slab-v7-regular.woff2 differ diff --git a/docs/_build/html/_static/fonts/fontawesome-webfont.eot b/docs/_build/html/_static/fonts/fontawesome-webfont.eot new file mode 100644 index 000000000..e9f60ca95 Binary files /dev/null and b/docs/_build/html/_static/fonts/fontawesome-webfont.eot differ diff --git a/docs/_build/html/_static/fonts/fontawesome-webfont.svg b/docs/_build/html/_static/fonts/fontawesome-webfont.svg new file mode 100644 index 000000000..855c845e5 --- /dev/null +++ b/docs/_build/html/_static/fonts/fontawesome-webfont.svg @@ -0,0 +1,2671 @@ + + + + +Created by FontForge 20120731 at Mon Oct 24 17:37:40 2016 + By ,,, +Copyright Dave Gandy 2016. All rights reserved. + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + diff --git a/docs/_build/html/_static/fonts/fontawesome-webfont.ttf b/docs/_build/html/_static/fonts/fontawesome-webfont.ttf new file mode 100644 index 000000000..35acda2fa Binary files /dev/null and b/docs/_build/html/_static/fonts/fontawesome-webfont.ttf differ diff --git a/docs/_build/html/_static/fonts/fontawesome-webfont.woff b/docs/_build/html/_static/fonts/fontawesome-webfont.woff new file mode 100644 index 000000000..400014a4b Binary files /dev/null and b/docs/_build/html/_static/fonts/fontawesome-webfont.woff differ diff --git a/docs/_build/html/_static/fonts/fontawesome-webfont.woff2 b/docs/_build/html/_static/fonts/fontawesome-webfont.woff2 new file mode 100644 index 000000000..4d13fc604 Binary files /dev/null and b/docs/_build/html/_static/fonts/fontawesome-webfont.woff2 differ diff --git a/docs/_build/html/_static/jquery-3.2.1.js b/docs/_build/html/_static/jquery-3.2.1.js new file mode 100644 index 000000000..d2d8ca479 --- /dev/null +++ b/docs/_build/html/_static/jquery-3.2.1.js @@ -0,0 +1,10253 @@ +/*! + * jQuery JavaScript Library v3.2.1 + * https://jquery.com/ + * + * Includes Sizzle.js + * https://sizzlejs.com/ + * + * Copyright JS Foundation and other contributors + * Released under the MIT license + * https://jquery.org/license + * + * Date: 2017-03-20T18:59Z + */ +( function( global, factory ) { + + "use strict"; + + if ( typeof module === "object" && typeof module.exports === "object" ) { + + // For CommonJS and CommonJS-like environments where a proper `window` + // is present, execute the factory and get jQuery. + // For environments that do not have a `window` with a `document` + // (such as Node.js), expose a factory as module.exports. + // This accentuates the need for the creation of a real `window`. + // e.g. var jQuery = require("jquery")(window); + // See ticket #14549 for more info. + module.exports = global.document ? + factory( global, true ) : + function( w ) { + if ( !w.document ) { + throw new Error( "jQuery requires a window with a document" ); + } + return factory( w ); + }; + } else { + factory( global ); + } + +// Pass this if window is not defined yet +} )( typeof window !== "undefined" ? window : this, function( window, noGlobal ) { + +// Edge <= 12 - 13+, Firefox <=18 - 45+, IE 10 - 11, Safari 5.1 - 9+, iOS 6 - 9.1 +// throw exceptions when non-strict code (e.g., ASP.NET 4.5) accesses strict mode +// arguments.callee.caller (trac-13335). But as of jQuery 3.0 (2016), strict mode should be common +// enough that all such attempts are guarded in a try block. +"use strict"; + +var arr = []; + +var document = window.document; + +var getProto = Object.getPrototypeOf; + +var slice = arr.slice; + +var concat = arr.concat; + +var push = arr.push; + +var indexOf = arr.indexOf; + +var class2type = {}; + +var toString = class2type.toString; + +var hasOwn = class2type.hasOwnProperty; + +var fnToString = hasOwn.toString; + +var ObjectFunctionString = fnToString.call( Object ); + +var support = {}; + + + + function DOMEval( code, doc ) { + doc = doc || document; + + var script = doc.createElement( "script" ); + + script.text = code; + doc.head.appendChild( script ).parentNode.removeChild( script ); + } +/* global Symbol */ +// Defining this global in .eslintrc.json would create a danger of using the global +// unguarded in another place, it seems safer to define global only for this module + + + +var + version = "3.2.1", + + // Define a local copy of jQuery + jQuery = function( selector, context ) { + + // The jQuery object is actually just the init constructor 'enhanced' + // Need init if jQuery is called (just allow error to be thrown if not included) + return new jQuery.fn.init( selector, context ); + }, + + // Support: Android <=4.0 only + // Make sure we trim BOM and NBSP + rtrim = /^[\s\uFEFF\xA0]+|[\s\uFEFF\xA0]+$/g, + + // Matches dashed string for camelizing + rmsPrefix = /^-ms-/, + rdashAlpha = /-([a-z])/g, + + // Used by jQuery.camelCase as callback to replace() + fcamelCase = function( all, letter ) { + return letter.toUpperCase(); + }; + +jQuery.fn = jQuery.prototype = { + + // The current version of jQuery being used + jquery: version, + + constructor: jQuery, + + // The default length of a jQuery object is 0 + length: 0, + + toArray: function() { + return slice.call( this ); + }, + + // Get the Nth element in the matched element set OR + // Get the whole matched element set as a clean array + get: function( num ) { + + // Return all the elements in a clean array + if ( num == null ) { + return slice.call( this ); + } + + // Return just the one element from the set + return num < 0 ? this[ num + this.length ] : this[ num ]; + }, + + // Take an array of elements and push it onto the stack + // (returning the new matched element set) + pushStack: function( elems ) { + + // Build a new jQuery matched element set + var ret = jQuery.merge( this.constructor(), elems ); + + // Add the old object onto the stack (as a reference) + ret.prevObject = this; + + // Return the newly-formed element set + return ret; + }, + + // Execute a callback for every element in the matched set. + each: function( callback ) { + return jQuery.each( this, callback ); + }, + + map: function( callback ) { + return this.pushStack( jQuery.map( this, function( elem, i ) { + return callback.call( elem, i, elem ); + } ) ); + }, + + slice: function() { + return this.pushStack( slice.apply( this, arguments ) ); + }, + + first: function() { + return this.eq( 0 ); + }, + + last: function() { + return this.eq( -1 ); + }, + + eq: function( i ) { + var len = this.length, + j = +i + ( i < 0 ? len : 0 ); + return this.pushStack( j >= 0 && j < len ? [ this[ j ] ] : [] ); + }, + + end: function() { + return this.prevObject || this.constructor(); + }, + + // For internal use only. + // Behaves like an Array's method, not like a jQuery method. + push: push, + sort: arr.sort, + splice: arr.splice +}; + +jQuery.extend = jQuery.fn.extend = function() { + var options, name, src, copy, copyIsArray, clone, + target = arguments[ 0 ] || {}, + i = 1, + length = arguments.length, + deep = false; + + // Handle a deep copy situation + if ( typeof target === "boolean" ) { + deep = target; + + // Skip the boolean and the target + target = arguments[ i ] || {}; + i++; + } + + // Handle case when target is a string or something (possible in deep copy) + if ( typeof target !== "object" && !jQuery.isFunction( target ) ) { + target = {}; + } + + // Extend jQuery itself if only one argument is passed + if ( i === length ) { + target = this; + i--; + } + + for ( ; i < length; i++ ) { + + // Only deal with non-null/undefined values + if ( ( options = arguments[ i ] ) != null ) { + + // Extend the base object + for ( name in options ) { + src = target[ name ]; + copy = options[ name ]; + + // Prevent never-ending loop + if ( target === copy ) { + continue; + } + + // Recurse if we're merging plain objects or arrays + if ( deep && copy && ( jQuery.isPlainObject( copy ) || + ( copyIsArray = Array.isArray( copy ) ) ) ) { + + if ( copyIsArray ) { + copyIsArray = false; + clone = src && Array.isArray( src ) ? src : []; + + } else { + clone = src && jQuery.isPlainObject( src ) ? src : {}; + } + + // Never move original objects, clone them + target[ name ] = jQuery.extend( deep, clone, copy ); + + // Don't bring in undefined values + } else if ( copy !== undefined ) { + target[ name ] = copy; + } + } + } + } + + // Return the modified object + return target; +}; + +jQuery.extend( { + + // Unique for each copy of jQuery on the page + expando: "jQuery" + ( version + Math.random() ).replace( /\D/g, "" ), + + // Assume jQuery is ready without the ready module + isReady: true, + + error: function( msg ) { + throw new Error( msg ); + }, + + noop: function() {}, + + isFunction: function( obj ) { + return jQuery.type( obj ) === "function"; + }, + + isWindow: function( obj ) { + return obj != null && obj === obj.window; + }, + + isNumeric: function( obj ) { + + // As of jQuery 3.0, isNumeric is limited to + // strings and numbers (primitives or objects) + // that can be coerced to finite numbers (gh-2662) + var type = jQuery.type( obj ); + return ( type === "number" || type === "string" ) && + + // parseFloat NaNs numeric-cast false positives ("") + // ...but misinterprets leading-number strings, particularly hex literals ("0x...") + // subtraction forces infinities to NaN + !isNaN( obj - parseFloat( obj ) ); + }, + + isPlainObject: function( obj ) { + var proto, Ctor; + + // Detect obvious negatives + // Use toString instead of jQuery.type to catch host objects + if ( !obj || toString.call( obj ) !== "[object Object]" ) { + return false; + } + + proto = getProto( obj ); + + // Objects with no prototype (e.g., `Object.create( null )`) are plain + if ( !proto ) { + return true; + } + + // Objects with prototype are plain iff they were constructed by a global Object function + Ctor = hasOwn.call( proto, "constructor" ) && proto.constructor; + return typeof Ctor === "function" && fnToString.call( Ctor ) === ObjectFunctionString; + }, + + isEmptyObject: function( obj ) { + + /* eslint-disable no-unused-vars */ + // See https://github.com/eslint/eslint/issues/6125 + var name; + + for ( name in obj ) { + return false; + } + return true; + }, + + type: function( obj ) { + if ( obj == null ) { + return obj + ""; + } + + // Support: Android <=2.3 only (functionish RegExp) + return typeof obj === "object" || typeof obj === "function" ? + class2type[ toString.call( obj ) ] || "object" : + typeof obj; + }, + + // Evaluates a script in a global context + globalEval: function( code ) { + DOMEval( code ); + }, + + // Convert dashed to camelCase; used by the css and data modules + // Support: IE <=9 - 11, Edge 12 - 13 + // Microsoft forgot to hump their vendor prefix (#9572) + camelCase: function( string ) { + return string.replace( rmsPrefix, "ms-" ).replace( rdashAlpha, fcamelCase ); + }, + + each: function( obj, callback ) { + var length, i = 0; + + if ( isArrayLike( obj ) ) { + length = obj.length; + for ( ; i < length; i++ ) { + if ( callback.call( obj[ i ], i, obj[ i ] ) === false ) { + break; + } + } + } else { + for ( i in obj ) { + if ( callback.call( obj[ i ], i, obj[ i ] ) === false ) { + break; + } + } + } + + return obj; + }, + + // Support: Android <=4.0 only + trim: function( text ) { + return text == null ? + "" : + ( text + "" ).replace( rtrim, "" ); + }, + + // results is for internal usage only + makeArray: function( arr, results ) { + var ret = results || []; + + if ( arr != null ) { + if ( isArrayLike( Object( arr ) ) ) { + jQuery.merge( ret, + typeof arr === "string" ? + [ arr ] : arr + ); + } else { + push.call( ret, arr ); + } + } + + return ret; + }, + + inArray: function( elem, arr, i ) { + return arr == null ? -1 : indexOf.call( arr, elem, i ); + }, + + // Support: Android <=4.0 only, PhantomJS 1 only + // push.apply(_, arraylike) throws on ancient WebKit + merge: function( first, second ) { + var len = +second.length, + j = 0, + i = first.length; + + for ( ; j < len; j++ ) { + first[ i++ ] = second[ j ]; + } + + first.length = i; + + return first; + }, + + grep: function( elems, callback, invert ) { + var callbackInverse, + matches = [], + i = 0, + length = elems.length, + callbackExpect = !invert; + + // Go through the array, only saving the items + // that pass the validator function + for ( ; i < length; i++ ) { + callbackInverse = !callback( elems[ i ], i ); + if ( callbackInverse !== callbackExpect ) { + matches.push( elems[ i ] ); + } + } + + return matches; + }, + + // arg is for internal usage only + map: function( elems, callback, arg ) { + var length, value, + i = 0, + ret = []; + + // Go through the array, translating each of the items to their new values + if ( isArrayLike( elems ) ) { + length = elems.length; + for ( ; i < length; i++ ) { + value = callback( elems[ i ], i, arg ); + + if ( value != null ) { + ret.push( value ); + } + } + + // Go through every key on the object, + } else { + for ( i in elems ) { + value = callback( elems[ i ], i, arg ); + + if ( value != null ) { + ret.push( value ); + } + } + } + + // Flatten any nested arrays + return concat.apply( [], ret ); + }, + + // A global GUID counter for objects + guid: 1, + + // Bind a function to a context, optionally partially applying any + // arguments. + proxy: function( fn, context ) { + var tmp, args, proxy; + + if ( typeof context === "string" ) { + tmp = fn[ context ]; + context = fn; + fn = tmp; + } + + // Quick check to determine if target is callable, in the spec + // this throws a TypeError, but we will just return undefined. + if ( !jQuery.isFunction( fn ) ) { + return undefined; + } + + // Simulated bind + args = slice.call( arguments, 2 ); + proxy = function() { + return fn.apply( context || this, args.concat( slice.call( arguments ) ) ); + }; + + // Set the guid of unique handler to the same of original handler, so it can be removed + proxy.guid = fn.guid = fn.guid || jQuery.guid++; + + return proxy; + }, + + now: Date.now, + + // jQuery.support is not used in Core but other projects attach their + // properties to it so it needs to exist. + support: support +} ); + +if ( typeof Symbol === "function" ) { + jQuery.fn[ Symbol.iterator ] = arr[ Symbol.iterator ]; +} + +// Populate the class2type map +jQuery.each( "Boolean Number String Function Array Date RegExp Object Error Symbol".split( " " ), +function( i, name ) { + class2type[ "[object " + name + "]" ] = name.toLowerCase(); +} ); + +function isArrayLike( obj ) { + + // Support: real iOS 8.2 only (not reproducible in simulator) + // `in` check used to prevent JIT error (gh-2145) + // hasOwn isn't used here due to false negatives + // regarding Nodelist length in IE + var length = !!obj && "length" in obj && obj.length, + type = jQuery.type( obj ); + + if ( type === "function" || jQuery.isWindow( obj ) ) { + return false; + } + + return type === "array" || length === 0 || + typeof length === "number" && length > 0 && ( length - 1 ) in obj; +} +var Sizzle = +/*! + * Sizzle CSS Selector Engine v2.3.3 + * https://sizzlejs.com/ + * + * Copyright jQuery Foundation and other contributors + * Released under the MIT license + * http://jquery.org/license + * + * Date: 2016-08-08 + */ +(function( window ) { + +var i, + support, + Expr, + getText, + isXML, + tokenize, + compile, + select, + outermostContext, + sortInput, + hasDuplicate, + + // Local document vars + setDocument, + document, + docElem, + documentIsHTML, + rbuggyQSA, + rbuggyMatches, + matches, + contains, + + // Instance-specific data + expando = "sizzle" + 1 * new Date(), + preferredDoc = window.document, + dirruns = 0, + done = 0, + classCache = createCache(), + tokenCache = createCache(), + compilerCache = createCache(), + sortOrder = function( a, b ) { + if ( a === b ) { + hasDuplicate = true; + } + return 0; + }, + + // Instance methods + hasOwn = ({}).hasOwnProperty, + arr = [], + pop = arr.pop, + push_native = arr.push, + push = arr.push, + slice = arr.slice, + // Use a stripped-down indexOf as it's faster than native + // https://jsperf.com/thor-indexof-vs-for/5 + indexOf = function( list, elem ) { + var i = 0, + len = list.length; + for ( ; i < len; i++ ) { + if ( list[i] === elem ) { + return i; + } + } + return -1; + }, + + booleans = "checked|selected|async|autofocus|autoplay|controls|defer|disabled|hidden|ismap|loop|multiple|open|readonly|required|scoped", + + // Regular expressions + + // http://www.w3.org/TR/css3-selectors/#whitespace + whitespace = "[\\x20\\t\\r\\n\\f]", + + // http://www.w3.org/TR/CSS21/syndata.html#value-def-identifier + identifier = "(?:\\\\.|[\\w-]|[^\0-\\xa0])+", + + // Attribute selectors: http://www.w3.org/TR/selectors/#attribute-selectors + attributes = "\\[" + whitespace + "*(" + identifier + ")(?:" + whitespace + + // Operator (capture 2) + "*([*^$|!~]?=)" + whitespace + + // "Attribute values must be CSS identifiers [capture 5] or strings [capture 3 or capture 4]" + "*(?:'((?:\\\\.|[^\\\\'])*)'|\"((?:\\\\.|[^\\\\\"])*)\"|(" + identifier + "))|)" + whitespace + + "*\\]", + + pseudos = ":(" + identifier + ")(?:\\((" + + // To reduce the number of selectors needing tokenize in the preFilter, prefer arguments: + // 1. quoted (capture 3; capture 4 or capture 5) + "('((?:\\\\.|[^\\\\'])*)'|\"((?:\\\\.|[^\\\\\"])*)\")|" + + // 2. simple (capture 6) + "((?:\\\\.|[^\\\\()[\\]]|" + attributes + ")*)|" + + // 3. anything else (capture 2) + ".*" + + ")\\)|)", + + // Leading and non-escaped trailing whitespace, capturing some non-whitespace characters preceding the latter + rwhitespace = new RegExp( whitespace + "+", "g" ), + rtrim = new RegExp( "^" + whitespace + "+|((?:^|[^\\\\])(?:\\\\.)*)" + whitespace + "+$", "g" ), + + rcomma = new RegExp( "^" + whitespace + "*," + whitespace + "*" ), + rcombinators = new RegExp( "^" + whitespace + "*([>+~]|" + whitespace + ")" + whitespace + "*" ), + + rattributeQuotes = new RegExp( "=" + whitespace + "*([^\\]'\"]*?)" + whitespace + "*\\]", "g" ), + + rpseudo = new RegExp( pseudos ), + ridentifier = new RegExp( "^" + identifier + "$" ), + + matchExpr = { + "ID": new RegExp( "^#(" + identifier + ")" ), + "CLASS": new RegExp( "^\\.(" + identifier + ")" ), + "TAG": new RegExp( "^(" + identifier + "|[*])" ), + "ATTR": new RegExp( "^" + attributes ), + "PSEUDO": new RegExp( "^" + pseudos ), + "CHILD": new RegExp( "^:(only|first|last|nth|nth-last)-(child|of-type)(?:\\(" + whitespace + + "*(even|odd|(([+-]|)(\\d*)n|)" + whitespace + "*(?:([+-]|)" + whitespace + + "*(\\d+)|))" + whitespace + "*\\)|)", "i" ), + "bool": new RegExp( "^(?:" + booleans + ")$", "i" ), + // For use in libraries implementing .is() + // We use this for POS matching in `select` + "needsContext": new RegExp( "^" + whitespace + "*[>+~]|:(even|odd|eq|gt|lt|nth|first|last)(?:\\(" + + whitespace + "*((?:-\\d)?\\d*)" + whitespace + "*\\)|)(?=[^-]|$)", "i" ) + }, + + rinputs = /^(?:input|select|textarea|button)$/i, + rheader = /^h\d$/i, + + rnative = /^[^{]+\{\s*\[native \w/, + + // Easily-parseable/retrievable ID or TAG or CLASS selectors + rquickExpr = /^(?:#([\w-]+)|(\w+)|\.([\w-]+))$/, + + rsibling = /[+~]/, + + // CSS escapes + // http://www.w3.org/TR/CSS21/syndata.html#escaped-characters + runescape = new RegExp( "\\\\([\\da-f]{1,6}" + whitespace + "?|(" + whitespace + ")|.)", "ig" ), + funescape = function( _, escaped, escapedWhitespace ) { + var high = "0x" + escaped - 0x10000; + // NaN means non-codepoint + // Support: Firefox<24 + // Workaround erroneous numeric interpretation of +"0x" + return high !== high || escapedWhitespace ? + escaped : + high < 0 ? + // BMP codepoint + String.fromCharCode( high + 0x10000 ) : + // Supplemental Plane codepoint (surrogate pair) + String.fromCharCode( high >> 10 | 0xD800, high & 0x3FF | 0xDC00 ); + }, + + // CSS string/identifier serialization + // https://drafts.csswg.org/cssom/#common-serializing-idioms + rcssescape = /([\0-\x1f\x7f]|^-?\d)|^-$|[^\0-\x1f\x7f-\uFFFF\w-]/g, + fcssescape = function( ch, asCodePoint ) { + if ( asCodePoint ) { + + // U+0000 NULL becomes U+FFFD REPLACEMENT CHARACTER + if ( ch === "\0" ) { + return "\uFFFD"; + } + + // Control characters and (dependent upon position) numbers get escaped as code points + return ch.slice( 0, -1 ) + "\\" + ch.charCodeAt( ch.length - 1 ).toString( 16 ) + " "; + } + + // Other potentially-special ASCII characters get backslash-escaped + return "\\" + ch; + }, + + // Used for iframes + // See setDocument() + // Removing the function wrapper causes a "Permission Denied" + // error in IE + unloadHandler = function() { + setDocument(); + }, + + disabledAncestor = addCombinator( + function( elem ) { + return elem.disabled === true && ("form" in elem || "label" in elem); + }, + { dir: "parentNode", next: "legend" } + ); + +// Optimize for push.apply( _, NodeList ) +try { + push.apply( + (arr = slice.call( preferredDoc.childNodes )), + preferredDoc.childNodes + ); + // Support: Android<4.0 + // Detect silently failing push.apply + arr[ preferredDoc.childNodes.length ].nodeType; +} catch ( e ) { + push = { apply: arr.length ? + + // Leverage slice if possible + function( target, els ) { + push_native.apply( target, slice.call(els) ); + } : + + // Support: IE<9 + // Otherwise append directly + function( target, els ) { + var j = target.length, + i = 0; + // Can't trust NodeList.length + while ( (target[j++] = els[i++]) ) {} + target.length = j - 1; + } + }; +} + +function Sizzle( selector, context, results, seed ) { + var m, i, elem, nid, match, groups, newSelector, + newContext = context && context.ownerDocument, + + // nodeType defaults to 9, since context defaults to document + nodeType = context ? context.nodeType : 9; + + results = results || []; + + // Return early from calls with invalid selector or context + if ( typeof selector !== "string" || !selector || + nodeType !== 1 && nodeType !== 9 && nodeType !== 11 ) { + + return results; + } + + // Try to shortcut find operations (as opposed to filters) in HTML documents + if ( !seed ) { + + if ( ( context ? context.ownerDocument || context : preferredDoc ) !== document ) { + setDocument( context ); + } + context = context || document; + + if ( documentIsHTML ) { + + // If the selector is sufficiently simple, try using a "get*By*" DOM method + // (excepting DocumentFragment context, where the methods don't exist) + if ( nodeType !== 11 && (match = rquickExpr.exec( selector )) ) { + + // ID selector + if ( (m = match[1]) ) { + + // Document context + if ( nodeType === 9 ) { + if ( (elem = context.getElementById( m )) ) { + + // Support: IE, Opera, Webkit + // TODO: identify versions + // getElementById can match elements by name instead of ID + if ( elem.id === m ) { + results.push( elem ); + return results; + } + } else { + return results; + } + + // Element context + } else { + + // Support: IE, Opera, Webkit + // TODO: identify versions + // getElementById can match elements by name instead of ID + if ( newContext && (elem = newContext.getElementById( m )) && + contains( context, elem ) && + elem.id === m ) { + + results.push( elem ); + return results; + } + } + + // Type selector + } else if ( match[2] ) { + push.apply( results, context.getElementsByTagName( selector ) ); + return results; + + // Class selector + } else if ( (m = match[3]) && support.getElementsByClassName && + context.getElementsByClassName ) { + + push.apply( results, context.getElementsByClassName( m ) ); + return results; + } + } + + // Take advantage of querySelectorAll + if ( support.qsa && + !compilerCache[ selector + " " ] && + (!rbuggyQSA || !rbuggyQSA.test( selector )) ) { + + if ( nodeType !== 1 ) { + newContext = context; + newSelector = selector; + + // qSA looks outside Element context, which is not what we want + // Thanks to Andrew Dupont for this workaround technique + // Support: IE <=8 + // Exclude object elements + } else if ( context.nodeName.toLowerCase() !== "object" ) { + + // Capture the context ID, setting it first if necessary + if ( (nid = context.getAttribute( "id" )) ) { + nid = nid.replace( rcssescape, fcssescape ); + } else { + context.setAttribute( "id", (nid = expando) ); + } + + // Prefix every selector in the list + groups = tokenize( selector ); + i = groups.length; + while ( i-- ) { + groups[i] = "#" + nid + " " + toSelector( groups[i] ); + } + newSelector = groups.join( "," ); + + // Expand context for sibling selectors + newContext = rsibling.test( selector ) && testContext( context.parentNode ) || + context; + } + + if ( newSelector ) { + try { + push.apply( results, + newContext.querySelectorAll( newSelector ) + ); + return results; + } catch ( qsaError ) { + } finally { + if ( nid === expando ) { + context.removeAttribute( "id" ); + } + } + } + } + } + } + + // All others + return select( selector.replace( rtrim, "$1" ), context, results, seed ); +} + +/** + * Create key-value caches of limited size + * @returns {function(string, object)} Returns the Object data after storing it on itself with + * property name the (space-suffixed) string and (if the cache is larger than Expr.cacheLength) + * deleting the oldest entry + */ +function createCache() { + var keys = []; + + function cache( key, value ) { + // Use (key + " ") to avoid collision with native prototype properties (see Issue #157) + if ( keys.push( key + " " ) > Expr.cacheLength ) { + // Only keep the most recent entries + delete cache[ keys.shift() ]; + } + return (cache[ key + " " ] = value); + } + return cache; +} + +/** + * Mark a function for special use by Sizzle + * @param {Function} fn The function to mark + */ +function markFunction( fn ) { + fn[ expando ] = true; + return fn; +} + +/** + * Support testing using an element + * @param {Function} fn Passed the created element and returns a boolean result + */ +function assert( fn ) { + var el = document.createElement("fieldset"); + + try { + return !!fn( el ); + } catch (e) { + return false; + } finally { + // Remove from its parent by default + if ( el.parentNode ) { + el.parentNode.removeChild( el ); + } + // release memory in IE + el = null; + } +} + +/** + * Adds the same handler for all of the specified attrs + * @param {String} attrs Pipe-separated list of attributes + * @param {Function} handler The method that will be applied + */ +function addHandle( attrs, handler ) { + var arr = attrs.split("|"), + i = arr.length; + + while ( i-- ) { + Expr.attrHandle[ arr[i] ] = handler; + } +} + +/** + * Checks document order of two siblings + * @param {Element} a + * @param {Element} b + * @returns {Number} Returns less than 0 if a precedes b, greater than 0 if a follows b + */ +function siblingCheck( a, b ) { + var cur = b && a, + diff = cur && a.nodeType === 1 && b.nodeType === 1 && + a.sourceIndex - b.sourceIndex; + + // Use IE sourceIndex if available on both nodes + if ( diff ) { + return diff; + } + + // Check if b follows a + if ( cur ) { + while ( (cur = cur.nextSibling) ) { + if ( cur === b ) { + return -1; + } + } + } + + return a ? 1 : -1; +} + +/** + * Returns a function to use in pseudos for input types + * @param {String} type + */ +function createInputPseudo( type ) { + return function( elem ) { + var name = elem.nodeName.toLowerCase(); + return name === "input" && elem.type === type; + }; +} + +/** + * Returns a function to use in pseudos for buttons + * @param {String} type + */ +function createButtonPseudo( type ) { + return function( elem ) { + var name = elem.nodeName.toLowerCase(); + return (name === "input" || name === "button") && elem.type === type; + }; +} + +/** + * Returns a function to use in pseudos for :enabled/:disabled + * @param {Boolean} disabled true for :disabled; false for :enabled + */ +function createDisabledPseudo( disabled ) { + + // Known :disabled false positives: fieldset[disabled] > legend:nth-of-type(n+2) :can-disable + return function( elem ) { + + // Only certain elements can match :enabled or :disabled + // https://html.spec.whatwg.org/multipage/scripting.html#selector-enabled + // https://html.spec.whatwg.org/multipage/scripting.html#selector-disabled + if ( "form" in elem ) { + + // Check for inherited disabledness on relevant non-disabled elements: + // * listed form-associated elements in a disabled fieldset + // https://html.spec.whatwg.org/multipage/forms.html#category-listed + // https://html.spec.whatwg.org/multipage/forms.html#concept-fe-disabled + // * option elements in a disabled optgroup + // https://html.spec.whatwg.org/multipage/forms.html#concept-option-disabled + // All such elements have a "form" property. + if ( elem.parentNode && elem.disabled === false ) { + + // Option elements defer to a parent optgroup if present + if ( "label" in elem ) { + if ( "label" in elem.parentNode ) { + return elem.parentNode.disabled === disabled; + } else { + return elem.disabled === disabled; + } + } + + // Support: IE 6 - 11 + // Use the isDisabled shortcut property to check for disabled fieldset ancestors + return elem.isDisabled === disabled || + + // Where there is no isDisabled, check manually + /* jshint -W018 */ + elem.isDisabled !== !disabled && + disabledAncestor( elem ) === disabled; + } + + return elem.disabled === disabled; + + // Try to winnow out elements that can't be disabled before trusting the disabled property. + // Some victims get caught in our net (label, legend, menu, track), but it shouldn't + // even exist on them, let alone have a boolean value. + } else if ( "label" in elem ) { + return elem.disabled === disabled; + } + + // Remaining elements are neither :enabled nor :disabled + return false; + }; +} + +/** + * Returns a function to use in pseudos for positionals + * @param {Function} fn + */ +function createPositionalPseudo( fn ) { + return markFunction(function( argument ) { + argument = +argument; + return markFunction(function( seed, matches ) { + var j, + matchIndexes = fn( [], seed.length, argument ), + i = matchIndexes.length; + + // Match elements found at the specified indexes + while ( i-- ) { + if ( seed[ (j = matchIndexes[i]) ] ) { + seed[j] = !(matches[j] = seed[j]); + } + } + }); + }); +} + +/** + * Checks a node for validity as a Sizzle context + * @param {Element|Object=} context + * @returns {Element|Object|Boolean} The input node if acceptable, otherwise a falsy value + */ +function testContext( context ) { + return context && typeof context.getElementsByTagName !== "undefined" && context; +} + +// Expose support vars for convenience +support = Sizzle.support = {}; + +/** + * Detects XML nodes + * @param {Element|Object} elem An element or a document + * @returns {Boolean} True iff elem is a non-HTML XML node + */ +isXML = Sizzle.isXML = function( elem ) { + // documentElement is verified for cases where it doesn't yet exist + // (such as loading iframes in IE - #4833) + var documentElement = elem && (elem.ownerDocument || elem).documentElement; + return documentElement ? documentElement.nodeName !== "HTML" : false; +}; + +/** + * Sets document-related variables once based on the current document + * @param {Element|Object} [doc] An element or document object to use to set the document + * @returns {Object} Returns the current document + */ +setDocument = Sizzle.setDocument = function( node ) { + var hasCompare, subWindow, + doc = node ? node.ownerDocument || node : preferredDoc; + + // Return early if doc is invalid or already selected + if ( doc === document || doc.nodeType !== 9 || !doc.documentElement ) { + return document; + } + + // Update global variables + document = doc; + docElem = document.documentElement; + documentIsHTML = !isXML( document ); + + // Support: IE 9-11, Edge + // Accessing iframe documents after unload throws "permission denied" errors (jQuery #13936) + if ( preferredDoc !== document && + (subWindow = document.defaultView) && subWindow.top !== subWindow ) { + + // Support: IE 11, Edge + if ( subWindow.addEventListener ) { + subWindow.addEventListener( "unload", unloadHandler, false ); + + // Support: IE 9 - 10 only + } else if ( subWindow.attachEvent ) { + subWindow.attachEvent( "onunload", unloadHandler ); + } + } + + /* Attributes + ---------------------------------------------------------------------- */ + + // Support: IE<8 + // Verify that getAttribute really returns attributes and not properties + // (excepting IE8 booleans) + support.attributes = assert(function( el ) { + el.className = "i"; + return !el.getAttribute("className"); + }); + + /* getElement(s)By* + ---------------------------------------------------------------------- */ + + // Check if getElementsByTagName("*") returns only elements + support.getElementsByTagName = assert(function( el ) { + el.appendChild( document.createComment("") ); + return !el.getElementsByTagName("*").length; + }); + + // Support: IE<9 + support.getElementsByClassName = rnative.test( document.getElementsByClassName ); + + // Support: IE<10 + // Check if getElementById returns elements by name + // The broken getElementById methods don't pick up programmatically-set names, + // so use a roundabout getElementsByName test + support.getById = assert(function( el ) { + docElem.appendChild( el ).id = expando; + return !document.getElementsByName || !document.getElementsByName( expando ).length; + }); + + // ID filter and find + if ( support.getById ) { + Expr.filter["ID"] = function( id ) { + var attrId = id.replace( runescape, funescape ); + return function( elem ) { + return elem.getAttribute("id") === attrId; + }; + }; + Expr.find["ID"] = function( id, context ) { + if ( typeof context.getElementById !== "undefined" && documentIsHTML ) { + var elem = context.getElementById( id ); + return elem ? [ elem ] : []; + } + }; + } else { + Expr.filter["ID"] = function( id ) { + var attrId = id.replace( runescape, funescape ); + return function( elem ) { + var node = typeof elem.getAttributeNode !== "undefined" && + elem.getAttributeNode("id"); + return node && node.value === attrId; + }; + }; + + // Support: IE 6 - 7 only + // getElementById is not reliable as a find shortcut + Expr.find["ID"] = function( id, context ) { + if ( typeof context.getElementById !== "undefined" && documentIsHTML ) { + var node, i, elems, + elem = context.getElementById( id ); + + if ( elem ) { + + // Verify the id attribute + node = elem.getAttributeNode("id"); + if ( node && node.value === id ) { + return [ elem ]; + } + + // Fall back on getElementsByName + elems = context.getElementsByName( id ); + i = 0; + while ( (elem = elems[i++]) ) { + node = elem.getAttributeNode("id"); + if ( node && node.value === id ) { + return [ elem ]; + } + } + } + + return []; + } + }; + } + + // Tag + Expr.find["TAG"] = support.getElementsByTagName ? + function( tag, context ) { + if ( typeof context.getElementsByTagName !== "undefined" ) { + return context.getElementsByTagName( tag ); + + // DocumentFragment nodes don't have gEBTN + } else if ( support.qsa ) { + return context.querySelectorAll( tag ); + } + } : + + function( tag, context ) { + var elem, + tmp = [], + i = 0, + // By happy coincidence, a (broken) gEBTN appears on DocumentFragment nodes too + results = context.getElementsByTagName( tag ); + + // Filter out possible comments + if ( tag === "*" ) { + while ( (elem = results[i++]) ) { + if ( elem.nodeType === 1 ) { + tmp.push( elem ); + } + } + + return tmp; + } + return results; + }; + + // Class + Expr.find["CLASS"] = support.getElementsByClassName && function( className, context ) { + if ( typeof context.getElementsByClassName !== "undefined" && documentIsHTML ) { + return context.getElementsByClassName( className ); + } + }; + + /* QSA/matchesSelector + ---------------------------------------------------------------------- */ + + // QSA and matchesSelector support + + // matchesSelector(:active) reports false when true (IE9/Opera 11.5) + rbuggyMatches = []; + + // qSa(:focus) reports false when true (Chrome 21) + // We allow this because of a bug in IE8/9 that throws an error + // whenever `document.activeElement` is accessed on an iframe + // So, we allow :focus to pass through QSA all the time to avoid the IE error + // See https://bugs.jquery.com/ticket/13378 + rbuggyQSA = []; + + if ( (support.qsa = rnative.test( document.querySelectorAll )) ) { + // Build QSA regex + // Regex strategy adopted from Diego Perini + assert(function( el ) { + // Select is set to empty string on purpose + // This is to test IE's treatment of not explicitly + // setting a boolean content attribute, + // since its presence should be enough + // https://bugs.jquery.com/ticket/12359 + docElem.appendChild( el ).innerHTML = "" + + ""; + + // Support: IE8, Opera 11-12.16 + // Nothing should be selected when empty strings follow ^= or $= or *= + // The test attribute must be unknown in Opera but "safe" for WinRT + // https://msdn.microsoft.com/en-us/library/ie/hh465388.aspx#attribute_section + if ( el.querySelectorAll("[msallowcapture^='']").length ) { + rbuggyQSA.push( "[*^$]=" + whitespace + "*(?:''|\"\")" ); + } + + // Support: IE8 + // Boolean attributes and "value" are not treated correctly + if ( !el.querySelectorAll("[selected]").length ) { + rbuggyQSA.push( "\\[" + whitespace + "*(?:value|" + booleans + ")" ); + } + + // Support: Chrome<29, Android<4.4, Safari<7.0+, iOS<7.0+, PhantomJS<1.9.8+ + if ( !el.querySelectorAll( "[id~=" + expando + "-]" ).length ) { + rbuggyQSA.push("~="); + } + + // Webkit/Opera - :checked should return selected option elements + // http://www.w3.org/TR/2011/REC-css3-selectors-20110929/#checked + // IE8 throws error here and will not see later tests + if ( !el.querySelectorAll(":checked").length ) { + rbuggyQSA.push(":checked"); + } + + // Support: Safari 8+, iOS 8+ + // https://bugs.webkit.org/show_bug.cgi?id=136851 + // In-page `selector#id sibling-combinator selector` fails + if ( !el.querySelectorAll( "a#" + expando + "+*" ).length ) { + rbuggyQSA.push(".#.+[+~]"); + } + }); + + assert(function( el ) { + el.innerHTML = "" + + ""; + + // Support: Windows 8 Native Apps + // The type and name attributes are restricted during .innerHTML assignment + var input = document.createElement("input"); + input.setAttribute( "type", "hidden" ); + el.appendChild( input ).setAttribute( "name", "D" ); + + // Support: IE8 + // Enforce case-sensitivity of name attribute + if ( el.querySelectorAll("[name=d]").length ) { + rbuggyQSA.push( "name" + whitespace + "*[*^$|!~]?=" ); + } + + // FF 3.5 - :enabled/:disabled and hidden elements (hidden elements are still enabled) + // IE8 throws error here and will not see later tests + if ( el.querySelectorAll(":enabled").length !== 2 ) { + rbuggyQSA.push( ":enabled", ":disabled" ); + } + + // Support: IE9-11+ + // IE's :disabled selector does not pick up the children of disabled fieldsets + docElem.appendChild( el ).disabled = true; + if ( el.querySelectorAll(":disabled").length !== 2 ) { + rbuggyQSA.push( ":enabled", ":disabled" ); + } + + // Opera 10-11 does not throw on post-comma invalid pseudos + el.querySelectorAll("*,:x"); + rbuggyQSA.push(",.*:"); + }); + } + + if ( (support.matchesSelector = rnative.test( (matches = docElem.matches || + docElem.webkitMatchesSelector || + docElem.mozMatchesSelector || + docElem.oMatchesSelector || + docElem.msMatchesSelector) )) ) { + + assert(function( el ) { + // Check to see if it's possible to do matchesSelector + // on a disconnected node (IE 9) + support.disconnectedMatch = matches.call( el, "*" ); + + // This should fail with an exception + // Gecko does not error, returns false instead + matches.call( el, "[s!='']:x" ); + rbuggyMatches.push( "!=", pseudos ); + }); + } + + rbuggyQSA = rbuggyQSA.length && new RegExp( rbuggyQSA.join("|") ); + rbuggyMatches = rbuggyMatches.length && new RegExp( rbuggyMatches.join("|") ); + + /* Contains + ---------------------------------------------------------------------- */ + hasCompare = rnative.test( docElem.compareDocumentPosition ); + + // Element contains another + // Purposefully self-exclusive + // As in, an element does not contain itself + contains = hasCompare || rnative.test( docElem.contains ) ? + function( a, b ) { + var adown = a.nodeType === 9 ? a.documentElement : a, + bup = b && b.parentNode; + return a === bup || !!( bup && bup.nodeType === 1 && ( + adown.contains ? + adown.contains( bup ) : + a.compareDocumentPosition && a.compareDocumentPosition( bup ) & 16 + )); + } : + function( a, b ) { + if ( b ) { + while ( (b = b.parentNode) ) { + if ( b === a ) { + return true; + } + } + } + return false; + }; + + /* Sorting + ---------------------------------------------------------------------- */ + + // Document order sorting + sortOrder = hasCompare ? + function( a, b ) { + + // Flag for duplicate removal + if ( a === b ) { + hasDuplicate = true; + return 0; + } + + // Sort on method existence if only one input has compareDocumentPosition + var compare = !a.compareDocumentPosition - !b.compareDocumentPosition; + if ( compare ) { + return compare; + } + + // Calculate position if both inputs belong to the same document + compare = ( a.ownerDocument || a ) === ( b.ownerDocument || b ) ? + a.compareDocumentPosition( b ) : + + // Otherwise we know they are disconnected + 1; + + // Disconnected nodes + if ( compare & 1 || + (!support.sortDetached && b.compareDocumentPosition( a ) === compare) ) { + + // Choose the first element that is related to our preferred document + if ( a === document || a.ownerDocument === preferredDoc && contains(preferredDoc, a) ) { + return -1; + } + if ( b === document || b.ownerDocument === preferredDoc && contains(preferredDoc, b) ) { + return 1; + } + + // Maintain original order + return sortInput ? + ( indexOf( sortInput, a ) - indexOf( sortInput, b ) ) : + 0; + } + + return compare & 4 ? -1 : 1; + } : + function( a, b ) { + // Exit early if the nodes are identical + if ( a === b ) { + hasDuplicate = true; + return 0; + } + + var cur, + i = 0, + aup = a.parentNode, + bup = b.parentNode, + ap = [ a ], + bp = [ b ]; + + // Parentless nodes are either documents or disconnected + if ( !aup || !bup ) { + return a === document ? -1 : + b === document ? 1 : + aup ? -1 : + bup ? 1 : + sortInput ? + ( indexOf( sortInput, a ) - indexOf( sortInput, b ) ) : + 0; + + // If the nodes are siblings, we can do a quick check + } else if ( aup === bup ) { + return siblingCheck( a, b ); + } + + // Otherwise we need full lists of their ancestors for comparison + cur = a; + while ( (cur = cur.parentNode) ) { + ap.unshift( cur ); + } + cur = b; + while ( (cur = cur.parentNode) ) { + bp.unshift( cur ); + } + + // Walk down the tree looking for a discrepancy + while ( ap[i] === bp[i] ) { + i++; + } + + return i ? + // Do a sibling check if the nodes have a common ancestor + siblingCheck( ap[i], bp[i] ) : + + // Otherwise nodes in our document sort first + ap[i] === preferredDoc ? -1 : + bp[i] === preferredDoc ? 1 : + 0; + }; + + return document; +}; + +Sizzle.matches = function( expr, elements ) { + return Sizzle( expr, null, null, elements ); +}; + +Sizzle.matchesSelector = function( elem, expr ) { + // Set document vars if needed + if ( ( elem.ownerDocument || elem ) !== document ) { + setDocument( elem ); + } + + // Make sure that attribute selectors are quoted + expr = expr.replace( rattributeQuotes, "='$1']" ); + + if ( support.matchesSelector && documentIsHTML && + !compilerCache[ expr + " " ] && + ( !rbuggyMatches || !rbuggyMatches.test( expr ) ) && + ( !rbuggyQSA || !rbuggyQSA.test( expr ) ) ) { + + try { + var ret = matches.call( elem, expr ); + + // IE 9's matchesSelector returns false on disconnected nodes + if ( ret || support.disconnectedMatch || + // As well, disconnected nodes are said to be in a document + // fragment in IE 9 + elem.document && elem.document.nodeType !== 11 ) { + return ret; + } + } catch (e) {} + } + + return Sizzle( expr, document, null, [ elem ] ).length > 0; +}; + +Sizzle.contains = function( context, elem ) { + // Set document vars if needed + if ( ( context.ownerDocument || context ) !== document ) { + setDocument( context ); + } + return contains( context, elem ); +}; + +Sizzle.attr = function( elem, name ) { + // Set document vars if needed + if ( ( elem.ownerDocument || elem ) !== document ) { + setDocument( elem ); + } + + var fn = Expr.attrHandle[ name.toLowerCase() ], + // Don't get fooled by Object.prototype properties (jQuery #13807) + val = fn && hasOwn.call( Expr.attrHandle, name.toLowerCase() ) ? + fn( elem, name, !documentIsHTML ) : + undefined; + + return val !== undefined ? + val : + support.attributes || !documentIsHTML ? + elem.getAttribute( name ) : + (val = elem.getAttributeNode(name)) && val.specified ? + val.value : + null; +}; + +Sizzle.escape = function( sel ) { + return (sel + "").replace( rcssescape, fcssescape ); +}; + +Sizzle.error = function( msg ) { + throw new Error( "Syntax error, unrecognized expression: " + msg ); +}; + +/** + * Document sorting and removing duplicates + * @param {ArrayLike} results + */ +Sizzle.uniqueSort = function( results ) { + var elem, + duplicates = [], + j = 0, + i = 0; + + // Unless we *know* we can detect duplicates, assume their presence + hasDuplicate = !support.detectDuplicates; + sortInput = !support.sortStable && results.slice( 0 ); + results.sort( sortOrder ); + + if ( hasDuplicate ) { + while ( (elem = results[i++]) ) { + if ( elem === results[ i ] ) { + j = duplicates.push( i ); + } + } + while ( j-- ) { + results.splice( duplicates[ j ], 1 ); + } + } + + // Clear input after sorting to release objects + // See https://github.com/jquery/sizzle/pull/225 + sortInput = null; + + return results; +}; + +/** + * Utility function for retrieving the text value of an array of DOM nodes + * @param {Array|Element} elem + */ +getText = Sizzle.getText = function( elem ) { + var node, + ret = "", + i = 0, + nodeType = elem.nodeType; + + if ( !nodeType ) { + // If no nodeType, this is expected to be an array + while ( (node = elem[i++]) ) { + // Do not traverse comment nodes + ret += getText( node ); + } + } else if ( nodeType === 1 || nodeType === 9 || nodeType === 11 ) { + // Use textContent for elements + // innerText usage removed for consistency of new lines (jQuery #11153) + if ( typeof elem.textContent === "string" ) { + return elem.textContent; + } else { + // Traverse its children + for ( elem = elem.firstChild; elem; elem = elem.nextSibling ) { + ret += getText( elem ); + } + } + } else if ( nodeType === 3 || nodeType === 4 ) { + return elem.nodeValue; + } + // Do not include comment or processing instruction nodes + + return ret; +}; + +Expr = Sizzle.selectors = { + + // Can be adjusted by the user + cacheLength: 50, + + createPseudo: markFunction, + + match: matchExpr, + + attrHandle: {}, + + find: {}, + + relative: { + ">": { dir: "parentNode", first: true }, + " ": { dir: "parentNode" }, + "+": { dir: "previousSibling", first: true }, + "~": { dir: "previousSibling" } + }, + + preFilter: { + "ATTR": function( match ) { + match[1] = match[1].replace( runescape, funescape ); + + // Move the given value to match[3] whether quoted or unquoted + match[3] = ( match[3] || match[4] || match[5] || "" ).replace( runescape, funescape ); + + if ( match[2] === "~=" ) { + match[3] = " " + match[3] + " "; + } + + return match.slice( 0, 4 ); + }, + + "CHILD": function( match ) { + /* matches from matchExpr["CHILD"] + 1 type (only|nth|...) + 2 what (child|of-type) + 3 argument (even|odd|\d*|\d*n([+-]\d+)?|...) + 4 xn-component of xn+y argument ([+-]?\d*n|) + 5 sign of xn-component + 6 x of xn-component + 7 sign of y-component + 8 y of y-component + */ + match[1] = match[1].toLowerCase(); + + if ( match[1].slice( 0, 3 ) === "nth" ) { + // nth-* requires argument + if ( !match[3] ) { + Sizzle.error( match[0] ); + } + + // numeric x and y parameters for Expr.filter.CHILD + // remember that false/true cast respectively to 0/1 + match[4] = +( match[4] ? match[5] + (match[6] || 1) : 2 * ( match[3] === "even" || match[3] === "odd" ) ); + match[5] = +( ( match[7] + match[8] ) || match[3] === "odd" ); + + // other types prohibit arguments + } else if ( match[3] ) { + Sizzle.error( match[0] ); + } + + return match; + }, + + "PSEUDO": function( match ) { + var excess, + unquoted = !match[6] && match[2]; + + if ( matchExpr["CHILD"].test( match[0] ) ) { + return null; + } + + // Accept quoted arguments as-is + if ( match[3] ) { + match[2] = match[4] || match[5] || ""; + + // Strip excess characters from unquoted arguments + } else if ( unquoted && rpseudo.test( unquoted ) && + // Get excess from tokenize (recursively) + (excess = tokenize( unquoted, true )) && + // advance to the next closing parenthesis + (excess = unquoted.indexOf( ")", unquoted.length - excess ) - unquoted.length) ) { + + // excess is a negative index + match[0] = match[0].slice( 0, excess ); + match[2] = unquoted.slice( 0, excess ); + } + + // Return only captures needed by the pseudo filter method (type and argument) + return match.slice( 0, 3 ); + } + }, + + filter: { + + "TAG": function( nodeNameSelector ) { + var nodeName = nodeNameSelector.replace( runescape, funescape ).toLowerCase(); + return nodeNameSelector === "*" ? + function() { return true; } : + function( elem ) { + return elem.nodeName && elem.nodeName.toLowerCase() === nodeName; + }; + }, + + "CLASS": function( className ) { + var pattern = classCache[ className + " " ]; + + return pattern || + (pattern = new RegExp( "(^|" + whitespace + ")" + className + "(" + whitespace + "|$)" )) && + classCache( className, function( elem ) { + return pattern.test( typeof elem.className === "string" && elem.className || typeof elem.getAttribute !== "undefined" && elem.getAttribute("class") || "" ); + }); + }, + + "ATTR": function( name, operator, check ) { + return function( elem ) { + var result = Sizzle.attr( elem, name ); + + if ( result == null ) { + return operator === "!="; + } + if ( !operator ) { + return true; + } + + result += ""; + + return operator === "=" ? result === check : + operator === "!=" ? result !== check : + operator === "^=" ? check && result.indexOf( check ) === 0 : + operator === "*=" ? check && result.indexOf( check ) > -1 : + operator === "$=" ? check && result.slice( -check.length ) === check : + operator === "~=" ? ( " " + result.replace( rwhitespace, " " ) + " " ).indexOf( check ) > -1 : + operator === "|=" ? result === check || result.slice( 0, check.length + 1 ) === check + "-" : + false; + }; + }, + + "CHILD": function( type, what, argument, first, last ) { + var simple = type.slice( 0, 3 ) !== "nth", + forward = type.slice( -4 ) !== "last", + ofType = what === "of-type"; + + return first === 1 && last === 0 ? + + // Shortcut for :nth-*(n) + function( elem ) { + return !!elem.parentNode; + } : + + function( elem, context, xml ) { + var cache, uniqueCache, outerCache, node, nodeIndex, start, + dir = simple !== forward ? "nextSibling" : "previousSibling", + parent = elem.parentNode, + name = ofType && elem.nodeName.toLowerCase(), + useCache = !xml && !ofType, + diff = false; + + if ( parent ) { + + // :(first|last|only)-(child|of-type) + if ( simple ) { + while ( dir ) { + node = elem; + while ( (node = node[ dir ]) ) { + if ( ofType ? + node.nodeName.toLowerCase() === name : + node.nodeType === 1 ) { + + return false; + } + } + // Reverse direction for :only-* (if we haven't yet done so) + start = dir = type === "only" && !start && "nextSibling"; + } + return true; + } + + start = [ forward ? parent.firstChild : parent.lastChild ]; + + // non-xml :nth-child(...) stores cache data on `parent` + if ( forward && useCache ) { + + // Seek `elem` from a previously-cached index + + // ...in a gzip-friendly way + node = parent; + outerCache = node[ expando ] || (node[ expando ] = {}); + + // Support: IE <9 only + // Defend against cloned attroperties (jQuery gh-1709) + uniqueCache = outerCache[ node.uniqueID ] || + (outerCache[ node.uniqueID ] = {}); + + cache = uniqueCache[ type ] || []; + nodeIndex = cache[ 0 ] === dirruns && cache[ 1 ]; + diff = nodeIndex && cache[ 2 ]; + node = nodeIndex && parent.childNodes[ nodeIndex ]; + + while ( (node = ++nodeIndex && node && node[ dir ] || + + // Fallback to seeking `elem` from the start + (diff = nodeIndex = 0) || start.pop()) ) { + + // When found, cache indexes on `parent` and break + if ( node.nodeType === 1 && ++diff && node === elem ) { + uniqueCache[ type ] = [ dirruns, nodeIndex, diff ]; + break; + } + } + + } else { + // Use previously-cached element index if available + if ( useCache ) { + // ...in a gzip-friendly way + node = elem; + outerCache = node[ expando ] || (node[ expando ] = {}); + + // Support: IE <9 only + // Defend against cloned attroperties (jQuery gh-1709) + uniqueCache = outerCache[ node.uniqueID ] || + (outerCache[ node.uniqueID ] = {}); + + cache = uniqueCache[ type ] || []; + nodeIndex = cache[ 0 ] === dirruns && cache[ 1 ]; + diff = nodeIndex; + } + + // xml :nth-child(...) + // or :nth-last-child(...) or :nth(-last)?-of-type(...) + if ( diff === false ) { + // Use the same loop as above to seek `elem` from the start + while ( (node = ++nodeIndex && node && node[ dir ] || + (diff = nodeIndex = 0) || start.pop()) ) { + + if ( ( ofType ? + node.nodeName.toLowerCase() === name : + node.nodeType === 1 ) && + ++diff ) { + + // Cache the index of each encountered element + if ( useCache ) { + outerCache = node[ expando ] || (node[ expando ] = {}); + + // Support: IE <9 only + // Defend against cloned attroperties (jQuery gh-1709) + uniqueCache = outerCache[ node.uniqueID ] || + (outerCache[ node.uniqueID ] = {}); + + uniqueCache[ type ] = [ dirruns, diff ]; + } + + if ( node === elem ) { + break; + } + } + } + } + } + + // Incorporate the offset, then check against cycle size + diff -= last; + return diff === first || ( diff % first === 0 && diff / first >= 0 ); + } + }; + }, + + "PSEUDO": function( pseudo, argument ) { + // pseudo-class names are case-insensitive + // http://www.w3.org/TR/selectors/#pseudo-classes + // Prioritize by case sensitivity in case custom pseudos are added with uppercase letters + // Remember that setFilters inherits from pseudos + var args, + fn = Expr.pseudos[ pseudo ] || Expr.setFilters[ pseudo.toLowerCase() ] || + Sizzle.error( "unsupported pseudo: " + pseudo ); + + // The user may use createPseudo to indicate that + // arguments are needed to create the filter function + // just as Sizzle does + if ( fn[ expando ] ) { + return fn( argument ); + } + + // But maintain support for old signatures + if ( fn.length > 1 ) { + args = [ pseudo, pseudo, "", argument ]; + return Expr.setFilters.hasOwnProperty( pseudo.toLowerCase() ) ? + markFunction(function( seed, matches ) { + var idx, + matched = fn( seed, argument ), + i = matched.length; + while ( i-- ) { + idx = indexOf( seed, matched[i] ); + seed[ idx ] = !( matches[ idx ] = matched[i] ); + } + }) : + function( elem ) { + return fn( elem, 0, args ); + }; + } + + return fn; + } + }, + + pseudos: { + // Potentially complex pseudos + "not": markFunction(function( selector ) { + // Trim the selector passed to compile + // to avoid treating leading and trailing + // spaces as combinators + var input = [], + results = [], + matcher = compile( selector.replace( rtrim, "$1" ) ); + + return matcher[ expando ] ? + markFunction(function( seed, matches, context, xml ) { + var elem, + unmatched = matcher( seed, null, xml, [] ), + i = seed.length; + + // Match elements unmatched by `matcher` + while ( i-- ) { + if ( (elem = unmatched[i]) ) { + seed[i] = !(matches[i] = elem); + } + } + }) : + function( elem, context, xml ) { + input[0] = elem; + matcher( input, null, xml, results ); + // Don't keep the element (issue #299) + input[0] = null; + return !results.pop(); + }; + }), + + "has": markFunction(function( selector ) { + return function( elem ) { + return Sizzle( selector, elem ).length > 0; + }; + }), + + "contains": markFunction(function( text ) { + text = text.replace( runescape, funescape ); + return function( elem ) { + return ( elem.textContent || elem.innerText || getText( elem ) ).indexOf( text ) > -1; + }; + }), + + // "Whether an element is represented by a :lang() selector + // is based solely on the element's language value + // being equal to the identifier C, + // or beginning with the identifier C immediately followed by "-". + // The matching of C against the element's language value is performed case-insensitively. + // The identifier C does not have to be a valid language name." + // http://www.w3.org/TR/selectors/#lang-pseudo + "lang": markFunction( function( lang ) { + // lang value must be a valid identifier + if ( !ridentifier.test(lang || "") ) { + Sizzle.error( "unsupported lang: " + lang ); + } + lang = lang.replace( runescape, funescape ).toLowerCase(); + return function( elem ) { + var elemLang; + do { + if ( (elemLang = documentIsHTML ? + elem.lang : + elem.getAttribute("xml:lang") || elem.getAttribute("lang")) ) { + + elemLang = elemLang.toLowerCase(); + return elemLang === lang || elemLang.indexOf( lang + "-" ) === 0; + } + } while ( (elem = elem.parentNode) && elem.nodeType === 1 ); + return false; + }; + }), + + // Miscellaneous + "target": function( elem ) { + var hash = window.location && window.location.hash; + return hash && hash.slice( 1 ) === elem.id; + }, + + "root": function( elem ) { + return elem === docElem; + }, + + "focus": function( elem ) { + return elem === document.activeElement && (!document.hasFocus || document.hasFocus()) && !!(elem.type || elem.href || ~elem.tabIndex); + }, + + // Boolean properties + "enabled": createDisabledPseudo( false ), + "disabled": createDisabledPseudo( true ), + + "checked": function( elem ) { + // In CSS3, :checked should return both checked and selected elements + // http://www.w3.org/TR/2011/REC-css3-selectors-20110929/#checked + var nodeName = elem.nodeName.toLowerCase(); + return (nodeName === "input" && !!elem.checked) || (nodeName === "option" && !!elem.selected); + }, + + "selected": function( elem ) { + // Accessing this property makes selected-by-default + // options in Safari work properly + if ( elem.parentNode ) { + elem.parentNode.selectedIndex; + } + + return elem.selected === true; + }, + + // Contents + "empty": function( elem ) { + // http://www.w3.org/TR/selectors/#empty-pseudo + // :empty is negated by element (1) or content nodes (text: 3; cdata: 4; entity ref: 5), + // but not by others (comment: 8; processing instruction: 7; etc.) + // nodeType < 6 works because attributes (2) do not appear as children + for ( elem = elem.firstChild; elem; elem = elem.nextSibling ) { + if ( elem.nodeType < 6 ) { + return false; + } + } + return true; + }, + + "parent": function( elem ) { + return !Expr.pseudos["empty"]( elem ); + }, + + // Element/input types + "header": function( elem ) { + return rheader.test( elem.nodeName ); + }, + + "input": function( elem ) { + return rinputs.test( elem.nodeName ); + }, + + "button": function( elem ) { + var name = elem.nodeName.toLowerCase(); + return name === "input" && elem.type === "button" || name === "button"; + }, + + "text": function( elem ) { + var attr; + return elem.nodeName.toLowerCase() === "input" && + elem.type === "text" && + + // Support: IE<8 + // New HTML5 attribute values (e.g., "search") appear with elem.type === "text" + ( (attr = elem.getAttribute("type")) == null || attr.toLowerCase() === "text" ); + }, + + // Position-in-collection + "first": createPositionalPseudo(function() { + return [ 0 ]; + }), + + "last": createPositionalPseudo(function( matchIndexes, length ) { + return [ length - 1 ]; + }), + + "eq": createPositionalPseudo(function( matchIndexes, length, argument ) { + return [ argument < 0 ? argument + length : argument ]; + }), + + "even": createPositionalPseudo(function( matchIndexes, length ) { + var i = 0; + for ( ; i < length; i += 2 ) { + matchIndexes.push( i ); + } + return matchIndexes; + }), + + "odd": createPositionalPseudo(function( matchIndexes, length ) { + var i = 1; + for ( ; i < length; i += 2 ) { + matchIndexes.push( i ); + } + return matchIndexes; + }), + + "lt": createPositionalPseudo(function( matchIndexes, length, argument ) { + var i = argument < 0 ? argument + length : argument; + for ( ; --i >= 0; ) { + matchIndexes.push( i ); + } + return matchIndexes; + }), + + "gt": createPositionalPseudo(function( matchIndexes, length, argument ) { + var i = argument < 0 ? argument + length : argument; + for ( ; ++i < length; ) { + matchIndexes.push( i ); + } + return matchIndexes; + }) + } +}; + +Expr.pseudos["nth"] = Expr.pseudos["eq"]; + +// Add button/input type pseudos +for ( i in { radio: true, checkbox: true, file: true, password: true, image: true } ) { + Expr.pseudos[ i ] = createInputPseudo( i ); +} +for ( i in { submit: true, reset: true } ) { + Expr.pseudos[ i ] = createButtonPseudo( i ); +} + +// Easy API for creating new setFilters +function setFilters() {} +setFilters.prototype = Expr.filters = Expr.pseudos; +Expr.setFilters = new setFilters(); + +tokenize = Sizzle.tokenize = function( selector, parseOnly ) { + var matched, match, tokens, type, + soFar, groups, preFilters, + cached = tokenCache[ selector + " " ]; + + if ( cached ) { + return parseOnly ? 0 : cached.slice( 0 ); + } + + soFar = selector; + groups = []; + preFilters = Expr.preFilter; + + while ( soFar ) { + + // Comma and first run + if ( !matched || (match = rcomma.exec( soFar )) ) { + if ( match ) { + // Don't consume trailing commas as valid + soFar = soFar.slice( match[0].length ) || soFar; + } + groups.push( (tokens = []) ); + } + + matched = false; + + // Combinators + if ( (match = rcombinators.exec( soFar )) ) { + matched = match.shift(); + tokens.push({ + value: matched, + // Cast descendant combinators to space + type: match[0].replace( rtrim, " " ) + }); + soFar = soFar.slice( matched.length ); + } + + // Filters + for ( type in Expr.filter ) { + if ( (match = matchExpr[ type ].exec( soFar )) && (!preFilters[ type ] || + (match = preFilters[ type ]( match ))) ) { + matched = match.shift(); + tokens.push({ + value: matched, + type: type, + matches: match + }); + soFar = soFar.slice( matched.length ); + } + } + + if ( !matched ) { + break; + } + } + + // Return the length of the invalid excess + // if we're just parsing + // Otherwise, throw an error or return tokens + return parseOnly ? + soFar.length : + soFar ? + Sizzle.error( selector ) : + // Cache the tokens + tokenCache( selector, groups ).slice( 0 ); +}; + +function toSelector( tokens ) { + var i = 0, + len = tokens.length, + selector = ""; + for ( ; i < len; i++ ) { + selector += tokens[i].value; + } + return selector; +} + +function addCombinator( matcher, combinator, base ) { + var dir = combinator.dir, + skip = combinator.next, + key = skip || dir, + checkNonElements = base && key === "parentNode", + doneName = done++; + + return combinator.first ? + // Check against closest ancestor/preceding element + function( elem, context, xml ) { + while ( (elem = elem[ dir ]) ) { + if ( elem.nodeType === 1 || checkNonElements ) { + return matcher( elem, context, xml ); + } + } + return false; + } : + + // Check against all ancestor/preceding elements + function( elem, context, xml ) { + var oldCache, uniqueCache, outerCache, + newCache = [ dirruns, doneName ]; + + // We can't set arbitrary data on XML nodes, so they don't benefit from combinator caching + if ( xml ) { + while ( (elem = elem[ dir ]) ) { + if ( elem.nodeType === 1 || checkNonElements ) { + if ( matcher( elem, context, xml ) ) { + return true; + } + } + } + } else { + while ( (elem = elem[ dir ]) ) { + if ( elem.nodeType === 1 || checkNonElements ) { + outerCache = elem[ expando ] || (elem[ expando ] = {}); + + // Support: IE <9 only + // Defend against cloned attroperties (jQuery gh-1709) + uniqueCache = outerCache[ elem.uniqueID ] || (outerCache[ elem.uniqueID ] = {}); + + if ( skip && skip === elem.nodeName.toLowerCase() ) { + elem = elem[ dir ] || elem; + } else if ( (oldCache = uniqueCache[ key ]) && + oldCache[ 0 ] === dirruns && oldCache[ 1 ] === doneName ) { + + // Assign to newCache so results back-propagate to previous elements + return (newCache[ 2 ] = oldCache[ 2 ]); + } else { + // Reuse newcache so results back-propagate to previous elements + uniqueCache[ key ] = newCache; + + // A match means we're done; a fail means we have to keep checking + if ( (newCache[ 2 ] = matcher( elem, context, xml )) ) { + return true; + } + } + } + } + } + return false; + }; +} + +function elementMatcher( matchers ) { + return matchers.length > 1 ? + function( elem, context, xml ) { + var i = matchers.length; + while ( i-- ) { + if ( !matchers[i]( elem, context, xml ) ) { + return false; + } + } + return true; + } : + matchers[0]; +} + +function multipleContexts( selector, contexts, results ) { + var i = 0, + len = contexts.length; + for ( ; i < len; i++ ) { + Sizzle( selector, contexts[i], results ); + } + return results; +} + +function condense( unmatched, map, filter, context, xml ) { + var elem, + newUnmatched = [], + i = 0, + len = unmatched.length, + mapped = map != null; + + for ( ; i < len; i++ ) { + if ( (elem = unmatched[i]) ) { + if ( !filter || filter( elem, context, xml ) ) { + newUnmatched.push( elem ); + if ( mapped ) { + map.push( i ); + } + } + } + } + + return newUnmatched; +} + +function setMatcher( preFilter, selector, matcher, postFilter, postFinder, postSelector ) { + if ( postFilter && !postFilter[ expando ] ) { + postFilter = setMatcher( postFilter ); + } + if ( postFinder && !postFinder[ expando ] ) { + postFinder = setMatcher( postFinder, postSelector ); + } + return markFunction(function( seed, results, context, xml ) { + var temp, i, elem, + preMap = [], + postMap = [], + preexisting = results.length, + + // Get initial elements from seed or context + elems = seed || multipleContexts( selector || "*", context.nodeType ? [ context ] : context, [] ), + + // Prefilter to get matcher input, preserving a map for seed-results synchronization + matcherIn = preFilter && ( seed || !selector ) ? + condense( elems, preMap, preFilter, context, xml ) : + elems, + + matcherOut = matcher ? + // If we have a postFinder, or filtered seed, or non-seed postFilter or preexisting results, + postFinder || ( seed ? preFilter : preexisting || postFilter ) ? + + // ...intermediate processing is necessary + [] : + + // ...otherwise use results directly + results : + matcherIn; + + // Find primary matches + if ( matcher ) { + matcher( matcherIn, matcherOut, context, xml ); + } + + // Apply postFilter + if ( postFilter ) { + temp = condense( matcherOut, postMap ); + postFilter( temp, [], context, xml ); + + // Un-match failing elements by moving them back to matcherIn + i = temp.length; + while ( i-- ) { + if ( (elem = temp[i]) ) { + matcherOut[ postMap[i] ] = !(matcherIn[ postMap[i] ] = elem); + } + } + } + + if ( seed ) { + if ( postFinder || preFilter ) { + if ( postFinder ) { + // Get the final matcherOut by condensing this intermediate into postFinder contexts + temp = []; + i = matcherOut.length; + while ( i-- ) { + if ( (elem = matcherOut[i]) ) { + // Restore matcherIn since elem is not yet a final match + temp.push( (matcherIn[i] = elem) ); + } + } + postFinder( null, (matcherOut = []), temp, xml ); + } + + // Move matched elements from seed to results to keep them synchronized + i = matcherOut.length; + while ( i-- ) { + if ( (elem = matcherOut[i]) && + (temp = postFinder ? indexOf( seed, elem ) : preMap[i]) > -1 ) { + + seed[temp] = !(results[temp] = elem); + } + } + } + + // Add elements to results, through postFinder if defined + } else { + matcherOut = condense( + matcherOut === results ? + matcherOut.splice( preexisting, matcherOut.length ) : + matcherOut + ); + if ( postFinder ) { + postFinder( null, results, matcherOut, xml ); + } else { + push.apply( results, matcherOut ); + } + } + }); +} + +function matcherFromTokens( tokens ) { + var checkContext, matcher, j, + len = tokens.length, + leadingRelative = Expr.relative[ tokens[0].type ], + implicitRelative = leadingRelative || Expr.relative[" "], + i = leadingRelative ? 1 : 0, + + // The foundational matcher ensures that elements are reachable from top-level context(s) + matchContext = addCombinator( function( elem ) { + return elem === checkContext; + }, implicitRelative, true ), + matchAnyContext = addCombinator( function( elem ) { + return indexOf( checkContext, elem ) > -1; + }, implicitRelative, true ), + matchers = [ function( elem, context, xml ) { + var ret = ( !leadingRelative && ( xml || context !== outermostContext ) ) || ( + (checkContext = context).nodeType ? + matchContext( elem, context, xml ) : + matchAnyContext( elem, context, xml ) ); + // Avoid hanging onto element (issue #299) + checkContext = null; + return ret; + } ]; + + for ( ; i < len; i++ ) { + if ( (matcher = Expr.relative[ tokens[i].type ]) ) { + matchers = [ addCombinator(elementMatcher( matchers ), matcher) ]; + } else { + matcher = Expr.filter[ tokens[i].type ].apply( null, tokens[i].matches ); + + // Return special upon seeing a positional matcher + if ( matcher[ expando ] ) { + // Find the next relative operator (if any) for proper handling + j = ++i; + for ( ; j < len; j++ ) { + if ( Expr.relative[ tokens[j].type ] ) { + break; + } + } + return setMatcher( + i > 1 && elementMatcher( matchers ), + i > 1 && toSelector( + // If the preceding token was a descendant combinator, insert an implicit any-element `*` + tokens.slice( 0, i - 1 ).concat({ value: tokens[ i - 2 ].type === " " ? "*" : "" }) + ).replace( rtrim, "$1" ), + matcher, + i < j && matcherFromTokens( tokens.slice( i, j ) ), + j < len && matcherFromTokens( (tokens = tokens.slice( j )) ), + j < len && toSelector( tokens ) + ); + } + matchers.push( matcher ); + } + } + + return elementMatcher( matchers ); +} + +function matcherFromGroupMatchers( elementMatchers, setMatchers ) { + var bySet = setMatchers.length > 0, + byElement = elementMatchers.length > 0, + superMatcher = function( seed, context, xml, results, outermost ) { + var elem, j, matcher, + matchedCount = 0, + i = "0", + unmatched = seed && [], + setMatched = [], + contextBackup = outermostContext, + // We must always have either seed elements or outermost context + elems = seed || byElement && Expr.find["TAG"]( "*", outermost ), + // Use integer dirruns iff this is the outermost matcher + dirrunsUnique = (dirruns += contextBackup == null ? 1 : Math.random() || 0.1), + len = elems.length; + + if ( outermost ) { + outermostContext = context === document || context || outermost; + } + + // Add elements passing elementMatchers directly to results + // Support: IE<9, Safari + // Tolerate NodeList properties (IE: "length"; Safari: ) matching elements by id + for ( ; i !== len && (elem = elems[i]) != null; i++ ) { + if ( byElement && elem ) { + j = 0; + if ( !context && elem.ownerDocument !== document ) { + setDocument( elem ); + xml = !documentIsHTML; + } + while ( (matcher = elementMatchers[j++]) ) { + if ( matcher( elem, context || document, xml) ) { + results.push( elem ); + break; + } + } + if ( outermost ) { + dirruns = dirrunsUnique; + } + } + + // Track unmatched elements for set filters + if ( bySet ) { + // They will have gone through all possible matchers + if ( (elem = !matcher && elem) ) { + matchedCount--; + } + + // Lengthen the array for every element, matched or not + if ( seed ) { + unmatched.push( elem ); + } + } + } + + // `i` is now the count of elements visited above, and adding it to `matchedCount` + // makes the latter nonnegative. + matchedCount += i; + + // Apply set filters to unmatched elements + // NOTE: This can be skipped if there are no unmatched elements (i.e., `matchedCount` + // equals `i`), unless we didn't visit _any_ elements in the above loop because we have + // no element matchers and no seed. + // Incrementing an initially-string "0" `i` allows `i` to remain a string only in that + // case, which will result in a "00" `matchedCount` that differs from `i` but is also + // numerically zero. + if ( bySet && i !== matchedCount ) { + j = 0; + while ( (matcher = setMatchers[j++]) ) { + matcher( unmatched, setMatched, context, xml ); + } + + if ( seed ) { + // Reintegrate element matches to eliminate the need for sorting + if ( matchedCount > 0 ) { + while ( i-- ) { + if ( !(unmatched[i] || setMatched[i]) ) { + setMatched[i] = pop.call( results ); + } + } + } + + // Discard index placeholder values to get only actual matches + setMatched = condense( setMatched ); + } + + // Add matches to results + push.apply( results, setMatched ); + + // Seedless set matches succeeding multiple successful matchers stipulate sorting + if ( outermost && !seed && setMatched.length > 0 && + ( matchedCount + setMatchers.length ) > 1 ) { + + Sizzle.uniqueSort( results ); + } + } + + // Override manipulation of globals by nested matchers + if ( outermost ) { + dirruns = dirrunsUnique; + outermostContext = contextBackup; + } + + return unmatched; + }; + + return bySet ? + markFunction( superMatcher ) : + superMatcher; +} + +compile = Sizzle.compile = function( selector, match /* Internal Use Only */ ) { + var i, + setMatchers = [], + elementMatchers = [], + cached = compilerCache[ selector + " " ]; + + if ( !cached ) { + // Generate a function of recursive functions that can be used to check each element + if ( !match ) { + match = tokenize( selector ); + } + i = match.length; + while ( i-- ) { + cached = matcherFromTokens( match[i] ); + if ( cached[ expando ] ) { + setMatchers.push( cached ); + } else { + elementMatchers.push( cached ); + } + } + + // Cache the compiled function + cached = compilerCache( selector, matcherFromGroupMatchers( elementMatchers, setMatchers ) ); + + // Save selector and tokenization + cached.selector = selector; + } + return cached; +}; + +/** + * A low-level selection function that works with Sizzle's compiled + * selector functions + * @param {String|Function} selector A selector or a pre-compiled + * selector function built with Sizzle.compile + * @param {Element} context + * @param {Array} [results] + * @param {Array} [seed] A set of elements to match against + */ +select = Sizzle.select = function( selector, context, results, seed ) { + var i, tokens, token, type, find, + compiled = typeof selector === "function" && selector, + match = !seed && tokenize( (selector = compiled.selector || selector) ); + + results = results || []; + + // Try to minimize operations if there is only one selector in the list and no seed + // (the latter of which guarantees us context) + if ( match.length === 1 ) { + + // Reduce context if the leading compound selector is an ID + tokens = match[0] = match[0].slice( 0 ); + if ( tokens.length > 2 && (token = tokens[0]).type === "ID" && + context.nodeType === 9 && documentIsHTML && Expr.relative[ tokens[1].type ] ) { + + context = ( Expr.find["ID"]( token.matches[0].replace(runescape, funescape), context ) || [] )[0]; + if ( !context ) { + return results; + + // Precompiled matchers will still verify ancestry, so step up a level + } else if ( compiled ) { + context = context.parentNode; + } + + selector = selector.slice( tokens.shift().value.length ); + } + + // Fetch a seed set for right-to-left matching + i = matchExpr["needsContext"].test( selector ) ? 0 : tokens.length; + while ( i-- ) { + token = tokens[i]; + + // Abort if we hit a combinator + if ( Expr.relative[ (type = token.type) ] ) { + break; + } + if ( (find = Expr.find[ type ]) ) { + // Search, expanding context for leading sibling combinators + if ( (seed = find( + token.matches[0].replace( runescape, funescape ), + rsibling.test( tokens[0].type ) && testContext( context.parentNode ) || context + )) ) { + + // If seed is empty or no tokens remain, we can return early + tokens.splice( i, 1 ); + selector = seed.length && toSelector( tokens ); + if ( !selector ) { + push.apply( results, seed ); + return results; + } + + break; + } + } + } + } + + // Compile and execute a filtering function if one is not provided + // Provide `match` to avoid retokenization if we modified the selector above + ( compiled || compile( selector, match ) )( + seed, + context, + !documentIsHTML, + results, + !context || rsibling.test( selector ) && testContext( context.parentNode ) || context + ); + return results; +}; + +// One-time assignments + +// Sort stability +support.sortStable = expando.split("").sort( sortOrder ).join("") === expando; + +// Support: Chrome 14-35+ +// Always assume duplicates if they aren't passed to the comparison function +support.detectDuplicates = !!hasDuplicate; + +// Initialize against the default document +setDocument(); + +// Support: Webkit<537.32 - Safari 6.0.3/Chrome 25 (fixed in Chrome 27) +// Detached nodes confoundingly follow *each other* +support.sortDetached = assert(function( el ) { + // Should return 1, but returns 4 (following) + return el.compareDocumentPosition( document.createElement("fieldset") ) & 1; +}); + +// Support: IE<8 +// Prevent attribute/property "interpolation" +// https://msdn.microsoft.com/en-us/library/ms536429%28VS.85%29.aspx +if ( !assert(function( el ) { + el.innerHTML = ""; + return el.firstChild.getAttribute("href") === "#" ; +}) ) { + addHandle( "type|href|height|width", function( elem, name, isXML ) { + if ( !isXML ) { + return elem.getAttribute( name, name.toLowerCase() === "type" ? 1 : 2 ); + } + }); +} + +// Support: IE<9 +// Use defaultValue in place of getAttribute("value") +if ( !support.attributes || !assert(function( el ) { + el.innerHTML = ""; + el.firstChild.setAttribute( "value", "" ); + return el.firstChild.getAttribute( "value" ) === ""; +}) ) { + addHandle( "value", function( elem, name, isXML ) { + if ( !isXML && elem.nodeName.toLowerCase() === "input" ) { + return elem.defaultValue; + } + }); +} + +// Support: IE<9 +// Use getAttributeNode to fetch booleans when getAttribute lies +if ( !assert(function( el ) { + return el.getAttribute("disabled") == null; +}) ) { + addHandle( booleans, function( elem, name, isXML ) { + var val; + if ( !isXML ) { + return elem[ name ] === true ? name.toLowerCase() : + (val = elem.getAttributeNode( name )) && val.specified ? + val.value : + null; + } + }); +} + +return Sizzle; + +})( window ); + + + +jQuery.find = Sizzle; +jQuery.expr = Sizzle.selectors; + +// Deprecated +jQuery.expr[ ":" ] = jQuery.expr.pseudos; +jQuery.uniqueSort = jQuery.unique = Sizzle.uniqueSort; +jQuery.text = Sizzle.getText; +jQuery.isXMLDoc = Sizzle.isXML; +jQuery.contains = Sizzle.contains; +jQuery.escapeSelector = Sizzle.escape; + + + + +var dir = function( elem, dir, until ) { + var matched = [], + truncate = until !== undefined; + + while ( ( elem = elem[ dir ] ) && elem.nodeType !== 9 ) { + if ( elem.nodeType === 1 ) { + if ( truncate && jQuery( elem ).is( until ) ) { + break; + } + matched.push( elem ); + } + } + return matched; +}; + + +var siblings = function( n, elem ) { + var matched = []; + + for ( ; n; n = n.nextSibling ) { + if ( n.nodeType === 1 && n !== elem ) { + matched.push( n ); + } + } + + return matched; +}; + + +var rneedsContext = jQuery.expr.match.needsContext; + + + +function nodeName( elem, name ) { + + return elem.nodeName && elem.nodeName.toLowerCase() === name.toLowerCase(); + +}; +var rsingleTag = ( /^<([a-z][^\/\0>:\x20\t\r\n\f]*)[\x20\t\r\n\f]*\/?>(?:<\/\1>|)$/i ); + + + +var risSimple = /^.[^:#\[\.,]*$/; + +// Implement the identical functionality for filter and not +function winnow( elements, qualifier, not ) { + if ( jQuery.isFunction( qualifier ) ) { + return jQuery.grep( elements, function( elem, i ) { + return !!qualifier.call( elem, i, elem ) !== not; + } ); + } + + // Single element + if ( qualifier.nodeType ) { + return jQuery.grep( elements, function( elem ) { + return ( elem === qualifier ) !== not; + } ); + } + + // Arraylike of elements (jQuery, arguments, Array) + if ( typeof qualifier !== "string" ) { + return jQuery.grep( elements, function( elem ) { + return ( indexOf.call( qualifier, elem ) > -1 ) !== not; + } ); + } + + // Simple selector that can be filtered directly, removing non-Elements + if ( risSimple.test( qualifier ) ) { + return jQuery.filter( qualifier, elements, not ); + } + + // Complex selector, compare the two sets, removing non-Elements + qualifier = jQuery.filter( qualifier, elements ); + return jQuery.grep( elements, function( elem ) { + return ( indexOf.call( qualifier, elem ) > -1 ) !== not && elem.nodeType === 1; + } ); +} + +jQuery.filter = function( expr, elems, not ) { + var elem = elems[ 0 ]; + + if ( not ) { + expr = ":not(" + expr + ")"; + } + + if ( elems.length === 1 && elem.nodeType === 1 ) { + return jQuery.find.matchesSelector( elem, expr ) ? [ elem ] : []; + } + + return jQuery.find.matches( expr, jQuery.grep( elems, function( elem ) { + return elem.nodeType === 1; + } ) ); +}; + +jQuery.fn.extend( { + find: function( selector ) { + var i, ret, + len = this.length, + self = this; + + if ( typeof selector !== "string" ) { + return this.pushStack( jQuery( selector ).filter( function() { + for ( i = 0; i < len; i++ ) { + if ( jQuery.contains( self[ i ], this ) ) { + return true; + } + } + } ) ); + } + + ret = this.pushStack( [] ); + + for ( i = 0; i < len; i++ ) { + jQuery.find( selector, self[ i ], ret ); + } + + return len > 1 ? jQuery.uniqueSort( ret ) : ret; + }, + filter: function( selector ) { + return this.pushStack( winnow( this, selector || [], false ) ); + }, + not: function( selector ) { + return this.pushStack( winnow( this, selector || [], true ) ); + }, + is: function( selector ) { + return !!winnow( + this, + + // If this is a positional/relative selector, check membership in the returned set + // so $("p:first").is("p:last") won't return true for a doc with two "p". + typeof selector === "string" && rneedsContext.test( selector ) ? + jQuery( selector ) : + selector || [], + false + ).length; + } +} ); + + +// Initialize a jQuery object + + +// A central reference to the root jQuery(document) +var rootjQuery, + + // A simple way to check for HTML strings + // Prioritize #id over to avoid XSS via location.hash (#9521) + // Strict HTML recognition (#11290: must start with <) + // Shortcut simple #id case for speed + rquickExpr = /^(?:\s*(<[\w\W]+>)[^>]*|#([\w-]+))$/, + + init = jQuery.fn.init = function( selector, context, root ) { + var match, elem; + + // HANDLE: $(""), $(null), $(undefined), $(false) + if ( !selector ) { + return this; + } + + // Method init() accepts an alternate rootjQuery + // so migrate can support jQuery.sub (gh-2101) + root = root || rootjQuery; + + // Handle HTML strings + if ( typeof selector === "string" ) { + if ( selector[ 0 ] === "<" && + selector[ selector.length - 1 ] === ">" && + selector.length >= 3 ) { + + // Assume that strings that start and end with <> are HTML and skip the regex check + match = [ null, selector, null ]; + + } else { + match = rquickExpr.exec( selector ); + } + + // Match html or make sure no context is specified for #id + if ( match && ( match[ 1 ] || !context ) ) { + + // HANDLE: $(html) -> $(array) + if ( match[ 1 ] ) { + context = context instanceof jQuery ? context[ 0 ] : context; + + // Option to run scripts is true for back-compat + // Intentionally let the error be thrown if parseHTML is not present + jQuery.merge( this, jQuery.parseHTML( + match[ 1 ], + context && context.nodeType ? context.ownerDocument || context : document, + true + ) ); + + // HANDLE: $(html, props) + if ( rsingleTag.test( match[ 1 ] ) && jQuery.isPlainObject( context ) ) { + for ( match in context ) { + + // Properties of context are called as methods if possible + if ( jQuery.isFunction( this[ match ] ) ) { + this[ match ]( context[ match ] ); + + // ...and otherwise set as attributes + } else { + this.attr( match, context[ match ] ); + } + } + } + + return this; + + // HANDLE: $(#id) + } else { + elem = document.getElementById( match[ 2 ] ); + + if ( elem ) { + + // Inject the element directly into the jQuery object + this[ 0 ] = elem; + this.length = 1; + } + return this; + } + + // HANDLE: $(expr, $(...)) + } else if ( !context || context.jquery ) { + return ( context || root ).find( selector ); + + // HANDLE: $(expr, context) + // (which is just equivalent to: $(context).find(expr) + } else { + return this.constructor( context ).find( selector ); + } + + // HANDLE: $(DOMElement) + } else if ( selector.nodeType ) { + this[ 0 ] = selector; + this.length = 1; + return this; + + // HANDLE: $(function) + // Shortcut for document ready + } else if ( jQuery.isFunction( selector ) ) { + return root.ready !== undefined ? + root.ready( selector ) : + + // Execute immediately if ready is not present + selector( jQuery ); + } + + return jQuery.makeArray( selector, this ); + }; + +// Give the init function the jQuery prototype for later instantiation +init.prototype = jQuery.fn; + +// Initialize central reference +rootjQuery = jQuery( document ); + + +var rparentsprev = /^(?:parents|prev(?:Until|All))/, + + // Methods guaranteed to produce a unique set when starting from a unique set + guaranteedUnique = { + children: true, + contents: true, + next: true, + prev: true + }; + +jQuery.fn.extend( { + has: function( target ) { + var targets = jQuery( target, this ), + l = targets.length; + + return this.filter( function() { + var i = 0; + for ( ; i < l; i++ ) { + if ( jQuery.contains( this, targets[ i ] ) ) { + return true; + } + } + } ); + }, + + closest: function( selectors, context ) { + var cur, + i = 0, + l = this.length, + matched = [], + targets = typeof selectors !== "string" && jQuery( selectors ); + + // Positional selectors never match, since there's no _selection_ context + if ( !rneedsContext.test( selectors ) ) { + for ( ; i < l; i++ ) { + for ( cur = this[ i ]; cur && cur !== context; cur = cur.parentNode ) { + + // Always skip document fragments + if ( cur.nodeType < 11 && ( targets ? + targets.index( cur ) > -1 : + + // Don't pass non-elements to Sizzle + cur.nodeType === 1 && + jQuery.find.matchesSelector( cur, selectors ) ) ) { + + matched.push( cur ); + break; + } + } + } + } + + return this.pushStack( matched.length > 1 ? jQuery.uniqueSort( matched ) : matched ); + }, + + // Determine the position of an element within the set + index: function( elem ) { + + // No argument, return index in parent + if ( !elem ) { + return ( this[ 0 ] && this[ 0 ].parentNode ) ? this.first().prevAll().length : -1; + } + + // Index in selector + if ( typeof elem === "string" ) { + return indexOf.call( jQuery( elem ), this[ 0 ] ); + } + + // Locate the position of the desired element + return indexOf.call( this, + + // If it receives a jQuery object, the first element is used + elem.jquery ? elem[ 0 ] : elem + ); + }, + + add: function( selector, context ) { + return this.pushStack( + jQuery.uniqueSort( + jQuery.merge( this.get(), jQuery( selector, context ) ) + ) + ); + }, + + addBack: function( selector ) { + return this.add( selector == null ? + this.prevObject : this.prevObject.filter( selector ) + ); + } +} ); + +function sibling( cur, dir ) { + while ( ( cur = cur[ dir ] ) && cur.nodeType !== 1 ) {} + return cur; +} + +jQuery.each( { + parent: function( elem ) { + var parent = elem.parentNode; + return parent && parent.nodeType !== 11 ? parent : null; + }, + parents: function( elem ) { + return dir( elem, "parentNode" ); + }, + parentsUntil: function( elem, i, until ) { + return dir( elem, "parentNode", until ); + }, + next: function( elem ) { + return sibling( elem, "nextSibling" ); + }, + prev: function( elem ) { + return sibling( elem, "previousSibling" ); + }, + nextAll: function( elem ) { + return dir( elem, "nextSibling" ); + }, + prevAll: function( elem ) { + return dir( elem, "previousSibling" ); + }, + nextUntil: function( elem, i, until ) { + return dir( elem, "nextSibling", until ); + }, + prevUntil: function( elem, i, until ) { + return dir( elem, "previousSibling", until ); + }, + siblings: function( elem ) { + return siblings( ( elem.parentNode || {} ).firstChild, elem ); + }, + children: function( elem ) { + return siblings( elem.firstChild ); + }, + contents: function( elem ) { + if ( nodeName( elem, "iframe" ) ) { + return elem.contentDocument; + } + + // Support: IE 9 - 11 only, iOS 7 only, Android Browser <=4.3 only + // Treat the template element as a regular one in browsers that + // don't support it. + if ( nodeName( elem, "template" ) ) { + elem = elem.content || elem; + } + + return jQuery.merge( [], elem.childNodes ); + } +}, function( name, fn ) { + jQuery.fn[ name ] = function( until, selector ) { + var matched = jQuery.map( this, fn, until ); + + if ( name.slice( -5 ) !== "Until" ) { + selector = until; + } + + if ( selector && typeof selector === "string" ) { + matched = jQuery.filter( selector, matched ); + } + + if ( this.length > 1 ) { + + // Remove duplicates + if ( !guaranteedUnique[ name ] ) { + jQuery.uniqueSort( matched ); + } + + // Reverse order for parents* and prev-derivatives + if ( rparentsprev.test( name ) ) { + matched.reverse(); + } + } + + return this.pushStack( matched ); + }; +} ); +var rnothtmlwhite = ( /[^\x20\t\r\n\f]+/g ); + + + +// Convert String-formatted options into Object-formatted ones +function createOptions( options ) { + var object = {}; + jQuery.each( options.match( rnothtmlwhite ) || [], function( _, flag ) { + object[ flag ] = true; + } ); + return object; +} + +/* + * Create a callback list using the following parameters: + * + * options: an optional list of space-separated options that will change how + * the callback list behaves or a more traditional option object + * + * By default a callback list will act like an event callback list and can be + * "fired" multiple times. + * + * Possible options: + * + * once: will ensure the callback list can only be fired once (like a Deferred) + * + * memory: will keep track of previous values and will call any callback added + * after the list has been fired right away with the latest "memorized" + * values (like a Deferred) + * + * unique: will ensure a callback can only be added once (no duplicate in the list) + * + * stopOnFalse: interrupt callings when a callback returns false + * + */ +jQuery.Callbacks = function( options ) { + + // Convert options from String-formatted to Object-formatted if needed + // (we check in cache first) + options = typeof options === "string" ? + createOptions( options ) : + jQuery.extend( {}, options ); + + var // Flag to know if list is currently firing + firing, + + // Last fire value for non-forgettable lists + memory, + + // Flag to know if list was already fired + fired, + + // Flag to prevent firing + locked, + + // Actual callback list + list = [], + + // Queue of execution data for repeatable lists + queue = [], + + // Index of currently firing callback (modified by add/remove as needed) + firingIndex = -1, + + // Fire callbacks + fire = function() { + + // Enforce single-firing + locked = locked || options.once; + + // Execute callbacks for all pending executions, + // respecting firingIndex overrides and runtime changes + fired = firing = true; + for ( ; queue.length; firingIndex = -1 ) { + memory = queue.shift(); + while ( ++firingIndex < list.length ) { + + // Run callback and check for early termination + if ( list[ firingIndex ].apply( memory[ 0 ], memory[ 1 ] ) === false && + options.stopOnFalse ) { + + // Jump to end and forget the data so .add doesn't re-fire + firingIndex = list.length; + memory = false; + } + } + } + + // Forget the data if we're done with it + if ( !options.memory ) { + memory = false; + } + + firing = false; + + // Clean up if we're done firing for good + if ( locked ) { + + // Keep an empty list if we have data for future add calls + if ( memory ) { + list = []; + + // Otherwise, this object is spent + } else { + list = ""; + } + } + }, + + // Actual Callbacks object + self = { + + // Add a callback or a collection of callbacks to the list + add: function() { + if ( list ) { + + // If we have memory from a past run, we should fire after adding + if ( memory && !firing ) { + firingIndex = list.length - 1; + queue.push( memory ); + } + + ( function add( args ) { + jQuery.each( args, function( _, arg ) { + if ( jQuery.isFunction( arg ) ) { + if ( !options.unique || !self.has( arg ) ) { + list.push( arg ); + } + } else if ( arg && arg.length && jQuery.type( arg ) !== "string" ) { + + // Inspect recursively + add( arg ); + } + } ); + } )( arguments ); + + if ( memory && !firing ) { + fire(); + } + } + return this; + }, + + // Remove a callback from the list + remove: function() { + jQuery.each( arguments, function( _, arg ) { + var index; + while ( ( index = jQuery.inArray( arg, list, index ) ) > -1 ) { + list.splice( index, 1 ); + + // Handle firing indexes + if ( index <= firingIndex ) { + firingIndex--; + } + } + } ); + return this; + }, + + // Check if a given callback is in the list. + // If no argument is given, return whether or not list has callbacks attached. + has: function( fn ) { + return fn ? + jQuery.inArray( fn, list ) > -1 : + list.length > 0; + }, + + // Remove all callbacks from the list + empty: function() { + if ( list ) { + list = []; + } + return this; + }, + + // Disable .fire and .add + // Abort any current/pending executions + // Clear all callbacks and values + disable: function() { + locked = queue = []; + list = memory = ""; + return this; + }, + disabled: function() { + return !list; + }, + + // Disable .fire + // Also disable .add unless we have memory (since it would have no effect) + // Abort any pending executions + lock: function() { + locked = queue = []; + if ( !memory && !firing ) { + list = memory = ""; + } + return this; + }, + locked: function() { + return !!locked; + }, + + // Call all callbacks with the given context and arguments + fireWith: function( context, args ) { + if ( !locked ) { + args = args || []; + args = [ context, args.slice ? args.slice() : args ]; + queue.push( args ); + if ( !firing ) { + fire(); + } + } + return this; + }, + + // Call all the callbacks with the given arguments + fire: function() { + self.fireWith( this, arguments ); + return this; + }, + + // To know if the callbacks have already been called at least once + fired: function() { + return !!fired; + } + }; + + return self; +}; + + +function Identity( v ) { + return v; +} +function Thrower( ex ) { + throw ex; +} + +function adoptValue( value, resolve, reject, noValue ) { + var method; + + try { + + // Check for promise aspect first to privilege synchronous behavior + if ( value && jQuery.isFunction( ( method = value.promise ) ) ) { + method.call( value ).done( resolve ).fail( reject ); + + // Other thenables + } else if ( value && jQuery.isFunction( ( method = value.then ) ) ) { + method.call( value, resolve, reject ); + + // Other non-thenables + } else { + + // Control `resolve` arguments by letting Array#slice cast boolean `noValue` to integer: + // * false: [ value ].slice( 0 ) => resolve( value ) + // * true: [ value ].slice( 1 ) => resolve() + resolve.apply( undefined, [ value ].slice( noValue ) ); + } + + // For Promises/A+, convert exceptions into rejections + // Since jQuery.when doesn't unwrap thenables, we can skip the extra checks appearing in + // Deferred#then to conditionally suppress rejection. + } catch ( value ) { + + // Support: Android 4.0 only + // Strict mode functions invoked without .call/.apply get global-object context + reject.apply( undefined, [ value ] ); + } +} + +jQuery.extend( { + + Deferred: function( func ) { + var tuples = [ + + // action, add listener, callbacks, + // ... .then handlers, argument index, [final state] + [ "notify", "progress", jQuery.Callbacks( "memory" ), + jQuery.Callbacks( "memory" ), 2 ], + [ "resolve", "done", jQuery.Callbacks( "once memory" ), + jQuery.Callbacks( "once memory" ), 0, "resolved" ], + [ "reject", "fail", jQuery.Callbacks( "once memory" ), + jQuery.Callbacks( "once memory" ), 1, "rejected" ] + ], + state = "pending", + promise = { + state: function() { + return state; + }, + always: function() { + deferred.done( arguments ).fail( arguments ); + return this; + }, + "catch": function( fn ) { + return promise.then( null, fn ); + }, + + // Keep pipe for back-compat + pipe: function( /* fnDone, fnFail, fnProgress */ ) { + var fns = arguments; + + return jQuery.Deferred( function( newDefer ) { + jQuery.each( tuples, function( i, tuple ) { + + // Map tuples (progress, done, fail) to arguments (done, fail, progress) + var fn = jQuery.isFunction( fns[ tuple[ 4 ] ] ) && fns[ tuple[ 4 ] ]; + + // deferred.progress(function() { bind to newDefer or newDefer.notify }) + // deferred.done(function() { bind to newDefer or newDefer.resolve }) + // deferred.fail(function() { bind to newDefer or newDefer.reject }) + deferred[ tuple[ 1 ] ]( function() { + var returned = fn && fn.apply( this, arguments ); + if ( returned && jQuery.isFunction( returned.promise ) ) { + returned.promise() + .progress( newDefer.notify ) + .done( newDefer.resolve ) + .fail( newDefer.reject ); + } else { + newDefer[ tuple[ 0 ] + "With" ]( + this, + fn ? [ returned ] : arguments + ); + } + } ); + } ); + fns = null; + } ).promise(); + }, + then: function( onFulfilled, onRejected, onProgress ) { + var maxDepth = 0; + function resolve( depth, deferred, handler, special ) { + return function() { + var that = this, + args = arguments, + mightThrow = function() { + var returned, then; + + // Support: Promises/A+ section 2.3.3.3.3 + // https://promisesaplus.com/#point-59 + // Ignore double-resolution attempts + if ( depth < maxDepth ) { + return; + } + + returned = handler.apply( that, args ); + + // Support: Promises/A+ section 2.3.1 + // https://promisesaplus.com/#point-48 + if ( returned === deferred.promise() ) { + throw new TypeError( "Thenable self-resolution" ); + } + + // Support: Promises/A+ sections 2.3.3.1, 3.5 + // https://promisesaplus.com/#point-54 + // https://promisesaplus.com/#point-75 + // Retrieve `then` only once + then = returned && + + // Support: Promises/A+ section 2.3.4 + // https://promisesaplus.com/#point-64 + // Only check objects and functions for thenability + ( typeof returned === "object" || + typeof returned === "function" ) && + returned.then; + + // Handle a returned thenable + if ( jQuery.isFunction( then ) ) { + + // Special processors (notify) just wait for resolution + if ( special ) { + then.call( + returned, + resolve( maxDepth, deferred, Identity, special ), + resolve( maxDepth, deferred, Thrower, special ) + ); + + // Normal processors (resolve) also hook into progress + } else { + + // ...and disregard older resolution values + maxDepth++; + + then.call( + returned, + resolve( maxDepth, deferred, Identity, special ), + resolve( maxDepth, deferred, Thrower, special ), + resolve( maxDepth, deferred, Identity, + deferred.notifyWith ) + ); + } + + // Handle all other returned values + } else { + + // Only substitute handlers pass on context + // and multiple values (non-spec behavior) + if ( handler !== Identity ) { + that = undefined; + args = [ returned ]; + } + + // Process the value(s) + // Default process is resolve + ( special || deferred.resolveWith )( that, args ); + } + }, + + // Only normal processors (resolve) catch and reject exceptions + process = special ? + mightThrow : + function() { + try { + mightThrow(); + } catch ( e ) { + + if ( jQuery.Deferred.exceptionHook ) { + jQuery.Deferred.exceptionHook( e, + process.stackTrace ); + } + + // Support: Promises/A+ section 2.3.3.3.4.1 + // https://promisesaplus.com/#point-61 + // Ignore post-resolution exceptions + if ( depth + 1 >= maxDepth ) { + + // Only substitute handlers pass on context + // and multiple values (non-spec behavior) + if ( handler !== Thrower ) { + that = undefined; + args = [ e ]; + } + + deferred.rejectWith( that, args ); + } + } + }; + + // Support: Promises/A+ section 2.3.3.3.1 + // https://promisesaplus.com/#point-57 + // Re-resolve promises immediately to dodge false rejection from + // subsequent errors + if ( depth ) { + process(); + } else { + + // Call an optional hook to record the stack, in case of exception + // since it's otherwise lost when execution goes async + if ( jQuery.Deferred.getStackHook ) { + process.stackTrace = jQuery.Deferred.getStackHook(); + } + window.setTimeout( process ); + } + }; + } + + return jQuery.Deferred( function( newDefer ) { + + // progress_handlers.add( ... ) + tuples[ 0 ][ 3 ].add( + resolve( + 0, + newDefer, + jQuery.isFunction( onProgress ) ? + onProgress : + Identity, + newDefer.notifyWith + ) + ); + + // fulfilled_handlers.add( ... ) + tuples[ 1 ][ 3 ].add( + resolve( + 0, + newDefer, + jQuery.isFunction( onFulfilled ) ? + onFulfilled : + Identity + ) + ); + + // rejected_handlers.add( ... ) + tuples[ 2 ][ 3 ].add( + resolve( + 0, + newDefer, + jQuery.isFunction( onRejected ) ? + onRejected : + Thrower + ) + ); + } ).promise(); + }, + + // Get a promise for this deferred + // If obj is provided, the promise aspect is added to the object + promise: function( obj ) { + return obj != null ? jQuery.extend( obj, promise ) : promise; + } + }, + deferred = {}; + + // Add list-specific methods + jQuery.each( tuples, function( i, tuple ) { + var list = tuple[ 2 ], + stateString = tuple[ 5 ]; + + // promise.progress = list.add + // promise.done = list.add + // promise.fail = list.add + promise[ tuple[ 1 ] ] = list.add; + + // Handle state + if ( stateString ) { + list.add( + function() { + + // state = "resolved" (i.e., fulfilled) + // state = "rejected" + state = stateString; + }, + + // rejected_callbacks.disable + // fulfilled_callbacks.disable + tuples[ 3 - i ][ 2 ].disable, + + // progress_callbacks.lock + tuples[ 0 ][ 2 ].lock + ); + } + + // progress_handlers.fire + // fulfilled_handlers.fire + // rejected_handlers.fire + list.add( tuple[ 3 ].fire ); + + // deferred.notify = function() { deferred.notifyWith(...) } + // deferred.resolve = function() { deferred.resolveWith(...) } + // deferred.reject = function() { deferred.rejectWith(...) } + deferred[ tuple[ 0 ] ] = function() { + deferred[ tuple[ 0 ] + "With" ]( this === deferred ? undefined : this, arguments ); + return this; + }; + + // deferred.notifyWith = list.fireWith + // deferred.resolveWith = list.fireWith + // deferred.rejectWith = list.fireWith + deferred[ tuple[ 0 ] + "With" ] = list.fireWith; + } ); + + // Make the deferred a promise + promise.promise( deferred ); + + // Call given func if any + if ( func ) { + func.call( deferred, deferred ); + } + + // All done! + return deferred; + }, + + // Deferred helper + when: function( singleValue ) { + var + + // count of uncompleted subordinates + remaining = arguments.length, + + // count of unprocessed arguments + i = remaining, + + // subordinate fulfillment data + resolveContexts = Array( i ), + resolveValues = slice.call( arguments ), + + // the master Deferred + master = jQuery.Deferred(), + + // subordinate callback factory + updateFunc = function( i ) { + return function( value ) { + resolveContexts[ i ] = this; + resolveValues[ i ] = arguments.length > 1 ? slice.call( arguments ) : value; + if ( !( --remaining ) ) { + master.resolveWith( resolveContexts, resolveValues ); + } + }; + }; + + // Single- and empty arguments are adopted like Promise.resolve + if ( remaining <= 1 ) { + adoptValue( singleValue, master.done( updateFunc( i ) ).resolve, master.reject, + !remaining ); + + // Use .then() to unwrap secondary thenables (cf. gh-3000) + if ( master.state() === "pending" || + jQuery.isFunction( resolveValues[ i ] && resolveValues[ i ].then ) ) { + + return master.then(); + } + } + + // Multiple arguments are aggregated like Promise.all array elements + while ( i-- ) { + adoptValue( resolveValues[ i ], updateFunc( i ), master.reject ); + } + + return master.promise(); + } +} ); + + +// These usually indicate a programmer mistake during development, +// warn about them ASAP rather than swallowing them by default. +var rerrorNames = /^(Eval|Internal|Range|Reference|Syntax|Type|URI)Error$/; + +jQuery.Deferred.exceptionHook = function( error, stack ) { + + // Support: IE 8 - 9 only + // Console exists when dev tools are open, which can happen at any time + if ( window.console && window.console.warn && error && rerrorNames.test( error.name ) ) { + window.console.warn( "jQuery.Deferred exception: " + error.message, error.stack, stack ); + } +}; + + + + +jQuery.readyException = function( error ) { + window.setTimeout( function() { + throw error; + } ); +}; + + + + +// The deferred used on DOM ready +var readyList = jQuery.Deferred(); + +jQuery.fn.ready = function( fn ) { + + readyList + .then( fn ) + + // Wrap jQuery.readyException in a function so that the lookup + // happens at the time of error handling instead of callback + // registration. + .catch( function( error ) { + jQuery.readyException( error ); + } ); + + return this; +}; + +jQuery.extend( { + + // Is the DOM ready to be used? Set to true once it occurs. + isReady: false, + + // A counter to track how many items to wait for before + // the ready event fires. See #6781 + readyWait: 1, + + // Handle when the DOM is ready + ready: function( wait ) { + + // Abort if there are pending holds or we're already ready + if ( wait === true ? --jQuery.readyWait : jQuery.isReady ) { + return; + } + + // Remember that the DOM is ready + jQuery.isReady = true; + + // If a normal DOM Ready event fired, decrement, and wait if need be + if ( wait !== true && --jQuery.readyWait > 0 ) { + return; + } + + // If there are functions bound, to execute + readyList.resolveWith( document, [ jQuery ] ); + } +} ); + +jQuery.ready.then = readyList.then; + +// The ready event handler and self cleanup method +function completed() { + document.removeEventListener( "DOMContentLoaded", completed ); + window.removeEventListener( "load", completed ); + jQuery.ready(); +} + +// Catch cases where $(document).ready() is called +// after the browser event has already occurred. +// Support: IE <=9 - 10 only +// Older IE sometimes signals "interactive" too soon +if ( document.readyState === "complete" || + ( document.readyState !== "loading" && !document.documentElement.doScroll ) ) { + + // Handle it asynchronously to allow scripts the opportunity to delay ready + window.setTimeout( jQuery.ready ); + +} else { + + // Use the handy event callback + document.addEventListener( "DOMContentLoaded", completed ); + + // A fallback to window.onload, that will always work + window.addEventListener( "load", completed ); +} + + + + +// Multifunctional method to get and set values of a collection +// The value/s can optionally be executed if it's a function +var access = function( elems, fn, key, value, chainable, emptyGet, raw ) { + var i = 0, + len = elems.length, + bulk = key == null; + + // Sets many values + if ( jQuery.type( key ) === "object" ) { + chainable = true; + for ( i in key ) { + access( elems, fn, i, key[ i ], true, emptyGet, raw ); + } + + // Sets one value + } else if ( value !== undefined ) { + chainable = true; + + if ( !jQuery.isFunction( value ) ) { + raw = true; + } + + if ( bulk ) { + + // Bulk operations run against the entire set + if ( raw ) { + fn.call( elems, value ); + fn = null; + + // ...except when executing function values + } else { + bulk = fn; + fn = function( elem, key, value ) { + return bulk.call( jQuery( elem ), value ); + }; + } + } + + if ( fn ) { + for ( ; i < len; i++ ) { + fn( + elems[ i ], key, raw ? + value : + value.call( elems[ i ], i, fn( elems[ i ], key ) ) + ); + } + } + } + + if ( chainable ) { + return elems; + } + + // Gets + if ( bulk ) { + return fn.call( elems ); + } + + return len ? fn( elems[ 0 ], key ) : emptyGet; +}; +var acceptData = function( owner ) { + + // Accepts only: + // - Node + // - Node.ELEMENT_NODE + // - Node.DOCUMENT_NODE + // - Object + // - Any + return owner.nodeType === 1 || owner.nodeType === 9 || !( +owner.nodeType ); +}; + + + + +function Data() { + this.expando = jQuery.expando + Data.uid++; +} + +Data.uid = 1; + +Data.prototype = { + + cache: function( owner ) { + + // Check if the owner object already has a cache + var value = owner[ this.expando ]; + + // If not, create one + if ( !value ) { + value = {}; + + // We can accept data for non-element nodes in modern browsers, + // but we should not, see #8335. + // Always return an empty object. + if ( acceptData( owner ) ) { + + // If it is a node unlikely to be stringify-ed or looped over + // use plain assignment + if ( owner.nodeType ) { + owner[ this.expando ] = value; + + // Otherwise secure it in a non-enumerable property + // configurable must be true to allow the property to be + // deleted when data is removed + } else { + Object.defineProperty( owner, this.expando, { + value: value, + configurable: true + } ); + } + } + } + + return value; + }, + set: function( owner, data, value ) { + var prop, + cache = this.cache( owner ); + + // Handle: [ owner, key, value ] args + // Always use camelCase key (gh-2257) + if ( typeof data === "string" ) { + cache[ jQuery.camelCase( data ) ] = value; + + // Handle: [ owner, { properties } ] args + } else { + + // Copy the properties one-by-one to the cache object + for ( prop in data ) { + cache[ jQuery.camelCase( prop ) ] = data[ prop ]; + } + } + return cache; + }, + get: function( owner, key ) { + return key === undefined ? + this.cache( owner ) : + + // Always use camelCase key (gh-2257) + owner[ this.expando ] && owner[ this.expando ][ jQuery.camelCase( key ) ]; + }, + access: function( owner, key, value ) { + + // In cases where either: + // + // 1. No key was specified + // 2. A string key was specified, but no value provided + // + // Take the "read" path and allow the get method to determine + // which value to return, respectively either: + // + // 1. The entire cache object + // 2. The data stored at the key + // + if ( key === undefined || + ( ( key && typeof key === "string" ) && value === undefined ) ) { + + return this.get( owner, key ); + } + + // When the key is not a string, or both a key and value + // are specified, set or extend (existing objects) with either: + // + // 1. An object of properties + // 2. A key and value + // + this.set( owner, key, value ); + + // Since the "set" path can have two possible entry points + // return the expected data based on which path was taken[*] + return value !== undefined ? value : key; + }, + remove: function( owner, key ) { + var i, + cache = owner[ this.expando ]; + + if ( cache === undefined ) { + return; + } + + if ( key !== undefined ) { + + // Support array or space separated string of keys + if ( Array.isArray( key ) ) { + + // If key is an array of keys... + // We always set camelCase keys, so remove that. + key = key.map( jQuery.camelCase ); + } else { + key = jQuery.camelCase( key ); + + // If a key with the spaces exists, use it. + // Otherwise, create an array by matching non-whitespace + key = key in cache ? + [ key ] : + ( key.match( rnothtmlwhite ) || [] ); + } + + i = key.length; + + while ( i-- ) { + delete cache[ key[ i ] ]; + } + } + + // Remove the expando if there's no more data + if ( key === undefined || jQuery.isEmptyObject( cache ) ) { + + // Support: Chrome <=35 - 45 + // Webkit & Blink performance suffers when deleting properties + // from DOM nodes, so set to undefined instead + // https://bugs.chromium.org/p/chromium/issues/detail?id=378607 (bug restricted) + if ( owner.nodeType ) { + owner[ this.expando ] = undefined; + } else { + delete owner[ this.expando ]; + } + } + }, + hasData: function( owner ) { + var cache = owner[ this.expando ]; + return cache !== undefined && !jQuery.isEmptyObject( cache ); + } +}; +var dataPriv = new Data(); + +var dataUser = new Data(); + + + +// Implementation Summary +// +// 1. Enforce API surface and semantic compatibility with 1.9.x branch +// 2. Improve the module's maintainability by reducing the storage +// paths to a single mechanism. +// 3. Use the same single mechanism to support "private" and "user" data. +// 4. _Never_ expose "private" data to user code (TODO: Drop _data, _removeData) +// 5. Avoid exposing implementation details on user objects (eg. expando properties) +// 6. Provide a clear path for implementation upgrade to WeakMap in 2014 + +var rbrace = /^(?:\{[\w\W]*\}|\[[\w\W]*\])$/, + rmultiDash = /[A-Z]/g; + +function getData( data ) { + if ( data === "true" ) { + return true; + } + + if ( data === "false" ) { + return false; + } + + if ( data === "null" ) { + return null; + } + + // Only convert to a number if it doesn't change the string + if ( data === +data + "" ) { + return +data; + } + + if ( rbrace.test( data ) ) { + return JSON.parse( data ); + } + + return data; +} + +function dataAttr( elem, key, data ) { + var name; + + // If nothing was found internally, try to fetch any + // data from the HTML5 data-* attribute + if ( data === undefined && elem.nodeType === 1 ) { + name = "data-" + key.replace( rmultiDash, "-$&" ).toLowerCase(); + data = elem.getAttribute( name ); + + if ( typeof data === "string" ) { + try { + data = getData( data ); + } catch ( e ) {} + + // Make sure we set the data so it isn't changed later + dataUser.set( elem, key, data ); + } else { + data = undefined; + } + } + return data; +} + +jQuery.extend( { + hasData: function( elem ) { + return dataUser.hasData( elem ) || dataPriv.hasData( elem ); + }, + + data: function( elem, name, data ) { + return dataUser.access( elem, name, data ); + }, + + removeData: function( elem, name ) { + dataUser.remove( elem, name ); + }, + + // TODO: Now that all calls to _data and _removeData have been replaced + // with direct calls to dataPriv methods, these can be deprecated. + _data: function( elem, name, data ) { + return dataPriv.access( elem, name, data ); + }, + + _removeData: function( elem, name ) { + dataPriv.remove( elem, name ); + } +} ); + +jQuery.fn.extend( { + data: function( key, value ) { + var i, name, data, + elem = this[ 0 ], + attrs = elem && elem.attributes; + + // Gets all values + if ( key === undefined ) { + if ( this.length ) { + data = dataUser.get( elem ); + + if ( elem.nodeType === 1 && !dataPriv.get( elem, "hasDataAttrs" ) ) { + i = attrs.length; + while ( i-- ) { + + // Support: IE 11 only + // The attrs elements can be null (#14894) + if ( attrs[ i ] ) { + name = attrs[ i ].name; + if ( name.indexOf( "data-" ) === 0 ) { + name = jQuery.camelCase( name.slice( 5 ) ); + dataAttr( elem, name, data[ name ] ); + } + } + } + dataPriv.set( elem, "hasDataAttrs", true ); + } + } + + return data; + } + + // Sets multiple values + if ( typeof key === "object" ) { + return this.each( function() { + dataUser.set( this, key ); + } ); + } + + return access( this, function( value ) { + var data; + + // The calling jQuery object (element matches) is not empty + // (and therefore has an element appears at this[ 0 ]) and the + // `value` parameter was not undefined. An empty jQuery object + // will result in `undefined` for elem = this[ 0 ] which will + // throw an exception if an attempt to read a data cache is made. + if ( elem && value === undefined ) { + + // Attempt to get data from the cache + // The key will always be camelCased in Data + data = dataUser.get( elem, key ); + if ( data !== undefined ) { + return data; + } + + // Attempt to "discover" the data in + // HTML5 custom data-* attrs + data = dataAttr( elem, key ); + if ( data !== undefined ) { + return data; + } + + // We tried really hard, but the data doesn't exist. + return; + } + + // Set the data... + this.each( function() { + + // We always store the camelCased key + dataUser.set( this, key, value ); + } ); + }, null, value, arguments.length > 1, null, true ); + }, + + removeData: function( key ) { + return this.each( function() { + dataUser.remove( this, key ); + } ); + } +} ); + + +jQuery.extend( { + queue: function( elem, type, data ) { + var queue; + + if ( elem ) { + type = ( type || "fx" ) + "queue"; + queue = dataPriv.get( elem, type ); + + // Speed up dequeue by getting out quickly if this is just a lookup + if ( data ) { + if ( !queue || Array.isArray( data ) ) { + queue = dataPriv.access( elem, type, jQuery.makeArray( data ) ); + } else { + queue.push( data ); + } + } + return queue || []; + } + }, + + dequeue: function( elem, type ) { + type = type || "fx"; + + var queue = jQuery.queue( elem, type ), + startLength = queue.length, + fn = queue.shift(), + hooks = jQuery._queueHooks( elem, type ), + next = function() { + jQuery.dequeue( elem, type ); + }; + + // If the fx queue is dequeued, always remove the progress sentinel + if ( fn === "inprogress" ) { + fn = queue.shift(); + startLength--; + } + + if ( fn ) { + + // Add a progress sentinel to prevent the fx queue from being + // automatically dequeued + if ( type === "fx" ) { + queue.unshift( "inprogress" ); + } + + // Clear up the last queue stop function + delete hooks.stop; + fn.call( elem, next, hooks ); + } + + if ( !startLength && hooks ) { + hooks.empty.fire(); + } + }, + + // Not public - generate a queueHooks object, or return the current one + _queueHooks: function( elem, type ) { + var key = type + "queueHooks"; + return dataPriv.get( elem, key ) || dataPriv.access( elem, key, { + empty: jQuery.Callbacks( "once memory" ).add( function() { + dataPriv.remove( elem, [ type + "queue", key ] ); + } ) + } ); + } +} ); + +jQuery.fn.extend( { + queue: function( type, data ) { + var setter = 2; + + if ( typeof type !== "string" ) { + data = type; + type = "fx"; + setter--; + } + + if ( arguments.length < setter ) { + return jQuery.queue( this[ 0 ], type ); + } + + return data === undefined ? + this : + this.each( function() { + var queue = jQuery.queue( this, type, data ); + + // Ensure a hooks for this queue + jQuery._queueHooks( this, type ); + + if ( type === "fx" && queue[ 0 ] !== "inprogress" ) { + jQuery.dequeue( this, type ); + } + } ); + }, + dequeue: function( type ) { + return this.each( function() { + jQuery.dequeue( this, type ); + } ); + }, + clearQueue: function( type ) { + return this.queue( type || "fx", [] ); + }, + + // Get a promise resolved when queues of a certain type + // are emptied (fx is the type by default) + promise: function( type, obj ) { + var tmp, + count = 1, + defer = jQuery.Deferred(), + elements = this, + i = this.length, + resolve = function() { + if ( !( --count ) ) { + defer.resolveWith( elements, [ elements ] ); + } + }; + + if ( typeof type !== "string" ) { + obj = type; + type = undefined; + } + type = type || "fx"; + + while ( i-- ) { + tmp = dataPriv.get( elements[ i ], type + "queueHooks" ); + if ( tmp && tmp.empty ) { + count++; + tmp.empty.add( resolve ); + } + } + resolve(); + return defer.promise( obj ); + } +} ); +var pnum = ( /[+-]?(?:\d*\.|)\d+(?:[eE][+-]?\d+|)/ ).source; + +var rcssNum = new RegExp( "^(?:([+-])=|)(" + pnum + ")([a-z%]*)$", "i" ); + + +var cssExpand = [ "Top", "Right", "Bottom", "Left" ]; + +var isHiddenWithinTree = function( elem, el ) { + + // isHiddenWithinTree might be called from jQuery#filter function; + // in that case, element will be second argument + elem = el || elem; + + // Inline style trumps all + return elem.style.display === "none" || + elem.style.display === "" && + + // Otherwise, check computed style + // Support: Firefox <=43 - 45 + // Disconnected elements can have computed display: none, so first confirm that elem is + // in the document. + jQuery.contains( elem.ownerDocument, elem ) && + + jQuery.css( elem, "display" ) === "none"; + }; + +var swap = function( elem, options, callback, args ) { + var ret, name, + old = {}; + + // Remember the old values, and insert the new ones + for ( name in options ) { + old[ name ] = elem.style[ name ]; + elem.style[ name ] = options[ name ]; + } + + ret = callback.apply( elem, args || [] ); + + // Revert the old values + for ( name in options ) { + elem.style[ name ] = old[ name ]; + } + + return ret; +}; + + + + +function adjustCSS( elem, prop, valueParts, tween ) { + var adjusted, + scale = 1, + maxIterations = 20, + currentValue = tween ? + function() { + return tween.cur(); + } : + function() { + return jQuery.css( elem, prop, "" ); + }, + initial = currentValue(), + unit = valueParts && valueParts[ 3 ] || ( jQuery.cssNumber[ prop ] ? "" : "px" ), + + // Starting value computation is required for potential unit mismatches + initialInUnit = ( jQuery.cssNumber[ prop ] || unit !== "px" && +initial ) && + rcssNum.exec( jQuery.css( elem, prop ) ); + + if ( initialInUnit && initialInUnit[ 3 ] !== unit ) { + + // Trust units reported by jQuery.css + unit = unit || initialInUnit[ 3 ]; + + // Make sure we update the tween properties later on + valueParts = valueParts || []; + + // Iteratively approximate from a nonzero starting point + initialInUnit = +initial || 1; + + do { + + // If previous iteration zeroed out, double until we get *something*. + // Use string for doubling so we don't accidentally see scale as unchanged below + scale = scale || ".5"; + + // Adjust and apply + initialInUnit = initialInUnit / scale; + jQuery.style( elem, prop, initialInUnit + unit ); + + // Update scale, tolerating zero or NaN from tween.cur() + // Break the loop if scale is unchanged or perfect, or if we've just had enough. + } while ( + scale !== ( scale = currentValue() / initial ) && scale !== 1 && --maxIterations + ); + } + + if ( valueParts ) { + initialInUnit = +initialInUnit || +initial || 0; + + // Apply relative offset (+=/-=) if specified + adjusted = valueParts[ 1 ] ? + initialInUnit + ( valueParts[ 1 ] + 1 ) * valueParts[ 2 ] : + +valueParts[ 2 ]; + if ( tween ) { + tween.unit = unit; + tween.start = initialInUnit; + tween.end = adjusted; + } + } + return adjusted; +} + + +var defaultDisplayMap = {}; + +function getDefaultDisplay( elem ) { + var temp, + doc = elem.ownerDocument, + nodeName = elem.nodeName, + display = defaultDisplayMap[ nodeName ]; + + if ( display ) { + return display; + } + + temp = doc.body.appendChild( doc.createElement( nodeName ) ); + display = jQuery.css( temp, "display" ); + + temp.parentNode.removeChild( temp ); + + if ( display === "none" ) { + display = "block"; + } + defaultDisplayMap[ nodeName ] = display; + + return display; +} + +function showHide( elements, show ) { + var display, elem, + values = [], + index = 0, + length = elements.length; + + // Determine new display value for elements that need to change + for ( ; index < length; index++ ) { + elem = elements[ index ]; + if ( !elem.style ) { + continue; + } + + display = elem.style.display; + if ( show ) { + + // Since we force visibility upon cascade-hidden elements, an immediate (and slow) + // check is required in this first loop unless we have a nonempty display value (either + // inline or about-to-be-restored) + if ( display === "none" ) { + values[ index ] = dataPriv.get( elem, "display" ) || null; + if ( !values[ index ] ) { + elem.style.display = ""; + } + } + if ( elem.style.display === "" && isHiddenWithinTree( elem ) ) { + values[ index ] = getDefaultDisplay( elem ); + } + } else { + if ( display !== "none" ) { + values[ index ] = "none"; + + // Remember what we're overwriting + dataPriv.set( elem, "display", display ); + } + } + } + + // Set the display of the elements in a second loop to avoid constant reflow + for ( index = 0; index < length; index++ ) { + if ( values[ index ] != null ) { + elements[ index ].style.display = values[ index ]; + } + } + + return elements; +} + +jQuery.fn.extend( { + show: function() { + return showHide( this, true ); + }, + hide: function() { + return showHide( this ); + }, + toggle: function( state ) { + if ( typeof state === "boolean" ) { + return state ? this.show() : this.hide(); + } + + return this.each( function() { + if ( isHiddenWithinTree( this ) ) { + jQuery( this ).show(); + } else { + jQuery( this ).hide(); + } + } ); + } +} ); +var rcheckableType = ( /^(?:checkbox|radio)$/i ); + +var rtagName = ( /<([a-z][^\/\0>\x20\t\r\n\f]+)/i ); + +var rscriptType = ( /^$|\/(?:java|ecma)script/i ); + + + +// We have to close these tags to support XHTML (#13200) +var wrapMap = { + + // Support: IE <=9 only + option: [ 1, "" ], + + // XHTML parsers do not magically insert elements in the + // same way that tag soup parsers do. So we cannot shorten + // this by omitting or other required elements. + thead: [ 1, "", "
" ], + col: [ 2, "", "
" ], + tr: [ 2, "", "
" ], + td: [ 3, "", "
" ], + + _default: [ 0, "", "" ] +}; + +// Support: IE <=9 only +wrapMap.optgroup = wrapMap.option; + +wrapMap.tbody = wrapMap.tfoot = wrapMap.colgroup = wrapMap.caption = wrapMap.thead; +wrapMap.th = wrapMap.td; + + +function getAll( context, tag ) { + + // Support: IE <=9 - 11 only + // Use typeof to avoid zero-argument method invocation on host objects (#15151) + var ret; + + if ( typeof context.getElementsByTagName !== "undefined" ) { + ret = context.getElementsByTagName( tag || "*" ); + + } else if ( typeof context.querySelectorAll !== "undefined" ) { + ret = context.querySelectorAll( tag || "*" ); + + } else { + ret = []; + } + + if ( tag === undefined || tag && nodeName( context, tag ) ) { + return jQuery.merge( [ context ], ret ); + } + + return ret; +} + + +// Mark scripts as having already been evaluated +function setGlobalEval( elems, refElements ) { + var i = 0, + l = elems.length; + + for ( ; i < l; i++ ) { + dataPriv.set( + elems[ i ], + "globalEval", + !refElements || dataPriv.get( refElements[ i ], "globalEval" ) + ); + } +} + + +var rhtml = /<|&#?\w+;/; + +function buildFragment( elems, context, scripts, selection, ignored ) { + var elem, tmp, tag, wrap, contains, j, + fragment = context.createDocumentFragment(), + nodes = [], + i = 0, + l = elems.length; + + for ( ; i < l; i++ ) { + elem = elems[ i ]; + + if ( elem || elem === 0 ) { + + // Add nodes directly + if ( jQuery.type( elem ) === "object" ) { + + // Support: Android <=4.0 only, PhantomJS 1 only + // push.apply(_, arraylike) throws on ancient WebKit + jQuery.merge( nodes, elem.nodeType ? [ elem ] : elem ); + + // Convert non-html into a text node + } else if ( !rhtml.test( elem ) ) { + nodes.push( context.createTextNode( elem ) ); + + // Convert html into DOM nodes + } else { + tmp = tmp || fragment.appendChild( context.createElement( "div" ) ); + + // Deserialize a standard representation + tag = ( rtagName.exec( elem ) || [ "", "" ] )[ 1 ].toLowerCase(); + wrap = wrapMap[ tag ] || wrapMap._default; + tmp.innerHTML = wrap[ 1 ] + jQuery.htmlPrefilter( elem ) + wrap[ 2 ]; + + // Descend through wrappers to the right content + j = wrap[ 0 ]; + while ( j-- ) { + tmp = tmp.lastChild; + } + + // Support: Android <=4.0 only, PhantomJS 1 only + // push.apply(_, arraylike) throws on ancient WebKit + jQuery.merge( nodes, tmp.childNodes ); + + // Remember the top-level container + tmp = fragment.firstChild; + + // Ensure the created nodes are orphaned (#12392) + tmp.textContent = ""; + } + } + } + + // Remove wrapper from fragment + fragment.textContent = ""; + + i = 0; + while ( ( elem = nodes[ i++ ] ) ) { + + // Skip elements already in the context collection (trac-4087) + if ( selection && jQuery.inArray( elem, selection ) > -1 ) { + if ( ignored ) { + ignored.push( elem ); + } + continue; + } + + contains = jQuery.contains( elem.ownerDocument, elem ); + + // Append to fragment + tmp = getAll( fragment.appendChild( elem ), "script" ); + + // Preserve script evaluation history + if ( contains ) { + setGlobalEval( tmp ); + } + + // Capture executables + if ( scripts ) { + j = 0; + while ( ( elem = tmp[ j++ ] ) ) { + if ( rscriptType.test( elem.type || "" ) ) { + scripts.push( elem ); + } + } + } + } + + return fragment; +} + + +( function() { + var fragment = document.createDocumentFragment(), + div = fragment.appendChild( document.createElement( "div" ) ), + input = document.createElement( "input" ); + + // Support: Android 4.0 - 4.3 only + // Check state lost if the name is set (#11217) + // Support: Windows Web Apps (WWA) + // `name` and `type` must use .setAttribute for WWA (#14901) + input.setAttribute( "type", "radio" ); + input.setAttribute( "checked", "checked" ); + input.setAttribute( "name", "t" ); + + div.appendChild( input ); + + // Support: Android <=4.1 only + // Older WebKit doesn't clone checked state correctly in fragments + support.checkClone = div.cloneNode( true ).cloneNode( true ).lastChild.checked; + + // Support: IE <=11 only + // Make sure textarea (and checkbox) defaultValue is properly cloned + div.innerHTML = ""; + support.noCloneChecked = !!div.cloneNode( true ).lastChild.defaultValue; +} )(); +var documentElement = document.documentElement; + + + +var + rkeyEvent = /^key/, + rmouseEvent = /^(?:mouse|pointer|contextmenu|drag|drop)|click/, + rtypenamespace = /^([^.]*)(?:\.(.+)|)/; + +function returnTrue() { + return true; +} + +function returnFalse() { + return false; +} + +// Support: IE <=9 only +// See #13393 for more info +function safeActiveElement() { + try { + return document.activeElement; + } catch ( err ) { } +} + +function on( elem, types, selector, data, fn, one ) { + var origFn, type; + + // Types can be a map of types/handlers + if ( typeof types === "object" ) { + + // ( types-Object, selector, data ) + if ( typeof selector !== "string" ) { + + // ( types-Object, data ) + data = data || selector; + selector = undefined; + } + for ( type in types ) { + on( elem, type, selector, data, types[ type ], one ); + } + return elem; + } + + if ( data == null && fn == null ) { + + // ( types, fn ) + fn = selector; + data = selector = undefined; + } else if ( fn == null ) { + if ( typeof selector === "string" ) { + + // ( types, selector, fn ) + fn = data; + data = undefined; + } else { + + // ( types, data, fn ) + fn = data; + data = selector; + selector = undefined; + } + } + if ( fn === false ) { + fn = returnFalse; + } else if ( !fn ) { + return elem; + } + + if ( one === 1 ) { + origFn = fn; + fn = function( event ) { + + // Can use an empty set, since event contains the info + jQuery().off( event ); + return origFn.apply( this, arguments ); + }; + + // Use same guid so caller can remove using origFn + fn.guid = origFn.guid || ( origFn.guid = jQuery.guid++ ); + } + return elem.each( function() { + jQuery.event.add( this, types, fn, data, selector ); + } ); +} + +/* + * Helper functions for managing events -- not part of the public interface. + * Props to Dean Edwards' addEvent library for many of the ideas. + */ +jQuery.event = { + + global: {}, + + add: function( elem, types, handler, data, selector ) { + + var handleObjIn, eventHandle, tmp, + events, t, handleObj, + special, handlers, type, namespaces, origType, + elemData = dataPriv.get( elem ); + + // Don't attach events to noData or text/comment nodes (but allow plain objects) + if ( !elemData ) { + return; + } + + // Caller can pass in an object of custom data in lieu of the handler + if ( handler.handler ) { + handleObjIn = handler; + handler = handleObjIn.handler; + selector = handleObjIn.selector; + } + + // Ensure that invalid selectors throw exceptions at attach time + // Evaluate against documentElement in case elem is a non-element node (e.g., document) + if ( selector ) { + jQuery.find.matchesSelector( documentElement, selector ); + } + + // Make sure that the handler has a unique ID, used to find/remove it later + if ( !handler.guid ) { + handler.guid = jQuery.guid++; + } + + // Init the element's event structure and main handler, if this is the first + if ( !( events = elemData.events ) ) { + events = elemData.events = {}; + } + if ( !( eventHandle = elemData.handle ) ) { + eventHandle = elemData.handle = function( e ) { + + // Discard the second event of a jQuery.event.trigger() and + // when an event is called after a page has unloaded + return typeof jQuery !== "undefined" && jQuery.event.triggered !== e.type ? + jQuery.event.dispatch.apply( elem, arguments ) : undefined; + }; + } + + // Handle multiple events separated by a space + types = ( types || "" ).match( rnothtmlwhite ) || [ "" ]; + t = types.length; + while ( t-- ) { + tmp = rtypenamespace.exec( types[ t ] ) || []; + type = origType = tmp[ 1 ]; + namespaces = ( tmp[ 2 ] || "" ).split( "." ).sort(); + + // There *must* be a type, no attaching namespace-only handlers + if ( !type ) { + continue; + } + + // If event changes its type, use the special event handlers for the changed type + special = jQuery.event.special[ type ] || {}; + + // If selector defined, determine special event api type, otherwise given type + type = ( selector ? special.delegateType : special.bindType ) || type; + + // Update special based on newly reset type + special = jQuery.event.special[ type ] || {}; + + // handleObj is passed to all event handlers + handleObj = jQuery.extend( { + type: type, + origType: origType, + data: data, + handler: handler, + guid: handler.guid, + selector: selector, + needsContext: selector && jQuery.expr.match.needsContext.test( selector ), + namespace: namespaces.join( "." ) + }, handleObjIn ); + + // Init the event handler queue if we're the first + if ( !( handlers = events[ type ] ) ) { + handlers = events[ type ] = []; + handlers.delegateCount = 0; + + // Only use addEventListener if the special events handler returns false + if ( !special.setup || + special.setup.call( elem, data, namespaces, eventHandle ) === false ) { + + if ( elem.addEventListener ) { + elem.addEventListener( type, eventHandle ); + } + } + } + + if ( special.add ) { + special.add.call( elem, handleObj ); + + if ( !handleObj.handler.guid ) { + handleObj.handler.guid = handler.guid; + } + } + + // Add to the element's handler list, delegates in front + if ( selector ) { + handlers.splice( handlers.delegateCount++, 0, handleObj ); + } else { + handlers.push( handleObj ); + } + + // Keep track of which events have ever been used, for event optimization + jQuery.event.global[ type ] = true; + } + + }, + + // Detach an event or set of events from an element + remove: function( elem, types, handler, selector, mappedTypes ) { + + var j, origCount, tmp, + events, t, handleObj, + special, handlers, type, namespaces, origType, + elemData = dataPriv.hasData( elem ) && dataPriv.get( elem ); + + if ( !elemData || !( events = elemData.events ) ) { + return; + } + + // Once for each type.namespace in types; type may be omitted + types = ( types || "" ).match( rnothtmlwhite ) || [ "" ]; + t = types.length; + while ( t-- ) { + tmp = rtypenamespace.exec( types[ t ] ) || []; + type = origType = tmp[ 1 ]; + namespaces = ( tmp[ 2 ] || "" ).split( "." ).sort(); + + // Unbind all events (on this namespace, if provided) for the element + if ( !type ) { + for ( type in events ) { + jQuery.event.remove( elem, type + types[ t ], handler, selector, true ); + } + continue; + } + + special = jQuery.event.special[ type ] || {}; + type = ( selector ? special.delegateType : special.bindType ) || type; + handlers = events[ type ] || []; + tmp = tmp[ 2 ] && + new RegExp( "(^|\\.)" + namespaces.join( "\\.(?:.*\\.|)" ) + "(\\.|$)" ); + + // Remove matching events + origCount = j = handlers.length; + while ( j-- ) { + handleObj = handlers[ j ]; + + if ( ( mappedTypes || origType === handleObj.origType ) && + ( !handler || handler.guid === handleObj.guid ) && + ( !tmp || tmp.test( handleObj.namespace ) ) && + ( !selector || selector === handleObj.selector || + selector === "**" && handleObj.selector ) ) { + handlers.splice( j, 1 ); + + if ( handleObj.selector ) { + handlers.delegateCount--; + } + if ( special.remove ) { + special.remove.call( elem, handleObj ); + } + } + } + + // Remove generic event handler if we removed something and no more handlers exist + // (avoids potential for endless recursion during removal of special event handlers) + if ( origCount && !handlers.length ) { + if ( !special.teardown || + special.teardown.call( elem, namespaces, elemData.handle ) === false ) { + + jQuery.removeEvent( elem, type, elemData.handle ); + } + + delete events[ type ]; + } + } + + // Remove data and the expando if it's no longer used + if ( jQuery.isEmptyObject( events ) ) { + dataPriv.remove( elem, "handle events" ); + } + }, + + dispatch: function( nativeEvent ) { + + // Make a writable jQuery.Event from the native event object + var event = jQuery.event.fix( nativeEvent ); + + var i, j, ret, matched, handleObj, handlerQueue, + args = new Array( arguments.length ), + handlers = ( dataPriv.get( this, "events" ) || {} )[ event.type ] || [], + special = jQuery.event.special[ event.type ] || {}; + + // Use the fix-ed jQuery.Event rather than the (read-only) native event + args[ 0 ] = event; + + for ( i = 1; i < arguments.length; i++ ) { + args[ i ] = arguments[ i ]; + } + + event.delegateTarget = this; + + // Call the preDispatch hook for the mapped type, and let it bail if desired + if ( special.preDispatch && special.preDispatch.call( this, event ) === false ) { + return; + } + + // Determine handlers + handlerQueue = jQuery.event.handlers.call( this, event, handlers ); + + // Run delegates first; they may want to stop propagation beneath us + i = 0; + while ( ( matched = handlerQueue[ i++ ] ) && !event.isPropagationStopped() ) { + event.currentTarget = matched.elem; + + j = 0; + while ( ( handleObj = matched.handlers[ j++ ] ) && + !event.isImmediatePropagationStopped() ) { + + // Triggered event must either 1) have no namespace, or 2) have namespace(s) + // a subset or equal to those in the bound event (both can have no namespace). + if ( !event.rnamespace || event.rnamespace.test( handleObj.namespace ) ) { + + event.handleObj = handleObj; + event.data = handleObj.data; + + ret = ( ( jQuery.event.special[ handleObj.origType ] || {} ).handle || + handleObj.handler ).apply( matched.elem, args ); + + if ( ret !== undefined ) { + if ( ( event.result = ret ) === false ) { + event.preventDefault(); + event.stopPropagation(); + } + } + } + } + } + + // Call the postDispatch hook for the mapped type + if ( special.postDispatch ) { + special.postDispatch.call( this, event ); + } + + return event.result; + }, + + handlers: function( event, handlers ) { + var i, handleObj, sel, matchedHandlers, matchedSelectors, + handlerQueue = [], + delegateCount = handlers.delegateCount, + cur = event.target; + + // Find delegate handlers + if ( delegateCount && + + // Support: IE <=9 + // Black-hole SVG instance trees (trac-13180) + cur.nodeType && + + // Support: Firefox <=42 + // Suppress spec-violating clicks indicating a non-primary pointer button (trac-3861) + // https://www.w3.org/TR/DOM-Level-3-Events/#event-type-click + // Support: IE 11 only + // ...but not arrow key "clicks" of radio inputs, which can have `button` -1 (gh-2343) + !( event.type === "click" && event.button >= 1 ) ) { + + for ( ; cur !== this; cur = cur.parentNode || this ) { + + // Don't check non-elements (#13208) + // Don't process clicks on disabled elements (#6911, #8165, #11382, #11764) + if ( cur.nodeType === 1 && !( event.type === "click" && cur.disabled === true ) ) { + matchedHandlers = []; + matchedSelectors = {}; + for ( i = 0; i < delegateCount; i++ ) { + handleObj = handlers[ i ]; + + // Don't conflict with Object.prototype properties (#13203) + sel = handleObj.selector + " "; + + if ( matchedSelectors[ sel ] === undefined ) { + matchedSelectors[ sel ] = handleObj.needsContext ? + jQuery( sel, this ).index( cur ) > -1 : + jQuery.find( sel, this, null, [ cur ] ).length; + } + if ( matchedSelectors[ sel ] ) { + matchedHandlers.push( handleObj ); + } + } + if ( matchedHandlers.length ) { + handlerQueue.push( { elem: cur, handlers: matchedHandlers } ); + } + } + } + } + + // Add the remaining (directly-bound) handlers + cur = this; + if ( delegateCount < handlers.length ) { + handlerQueue.push( { elem: cur, handlers: handlers.slice( delegateCount ) } ); + } + + return handlerQueue; + }, + + addProp: function( name, hook ) { + Object.defineProperty( jQuery.Event.prototype, name, { + enumerable: true, + configurable: true, + + get: jQuery.isFunction( hook ) ? + function() { + if ( this.originalEvent ) { + return hook( this.originalEvent ); + } + } : + function() { + if ( this.originalEvent ) { + return this.originalEvent[ name ]; + } + }, + + set: function( value ) { + Object.defineProperty( this, name, { + enumerable: true, + configurable: true, + writable: true, + value: value + } ); + } + } ); + }, + + fix: function( originalEvent ) { + return originalEvent[ jQuery.expando ] ? + originalEvent : + new jQuery.Event( originalEvent ); + }, + + special: { + load: { + + // Prevent triggered image.load events from bubbling to window.load + noBubble: true + }, + focus: { + + // Fire native event if possible so blur/focus sequence is correct + trigger: function() { + if ( this !== safeActiveElement() && this.focus ) { + this.focus(); + return false; + } + }, + delegateType: "focusin" + }, + blur: { + trigger: function() { + if ( this === safeActiveElement() && this.blur ) { + this.blur(); + return false; + } + }, + delegateType: "focusout" + }, + click: { + + // For checkbox, fire native event so checked state will be right + trigger: function() { + if ( this.type === "checkbox" && this.click && nodeName( this, "input" ) ) { + this.click(); + return false; + } + }, + + // For cross-browser consistency, don't fire native .click() on links + _default: function( event ) { + return nodeName( event.target, "a" ); + } + }, + + beforeunload: { + postDispatch: function( event ) { + + // Support: Firefox 20+ + // Firefox doesn't alert if the returnValue field is not set. + if ( event.result !== undefined && event.originalEvent ) { + event.originalEvent.returnValue = event.result; + } + } + } + } +}; + +jQuery.removeEvent = function( elem, type, handle ) { + + // This "if" is needed for plain objects + if ( elem.removeEventListener ) { + elem.removeEventListener( type, handle ); + } +}; + +jQuery.Event = function( src, props ) { + + // Allow instantiation without the 'new' keyword + if ( !( this instanceof jQuery.Event ) ) { + return new jQuery.Event( src, props ); + } + + // Event object + if ( src && src.type ) { + this.originalEvent = src; + this.type = src.type; + + // Events bubbling up the document may have been marked as prevented + // by a handler lower down the tree; reflect the correct value. + this.isDefaultPrevented = src.defaultPrevented || + src.defaultPrevented === undefined && + + // Support: Android <=2.3 only + src.returnValue === false ? + returnTrue : + returnFalse; + + // Create target properties + // Support: Safari <=6 - 7 only + // Target should not be a text node (#504, #13143) + this.target = ( src.target && src.target.nodeType === 3 ) ? + src.target.parentNode : + src.target; + + this.currentTarget = src.currentTarget; + this.relatedTarget = src.relatedTarget; + + // Event type + } else { + this.type = src; + } + + // Put explicitly provided properties onto the event object + if ( props ) { + jQuery.extend( this, props ); + } + + // Create a timestamp if incoming event doesn't have one + this.timeStamp = src && src.timeStamp || jQuery.now(); + + // Mark it as fixed + this[ jQuery.expando ] = true; +}; + +// jQuery.Event is based on DOM3 Events as specified by the ECMAScript Language Binding +// https://www.w3.org/TR/2003/WD-DOM-Level-3-Events-20030331/ecma-script-binding.html +jQuery.Event.prototype = { + constructor: jQuery.Event, + isDefaultPrevented: returnFalse, + isPropagationStopped: returnFalse, + isImmediatePropagationStopped: returnFalse, + isSimulated: false, + + preventDefault: function() { + var e = this.originalEvent; + + this.isDefaultPrevented = returnTrue; + + if ( e && !this.isSimulated ) { + e.preventDefault(); + } + }, + stopPropagation: function() { + var e = this.originalEvent; + + this.isPropagationStopped = returnTrue; + + if ( e && !this.isSimulated ) { + e.stopPropagation(); + } + }, + stopImmediatePropagation: function() { + var e = this.originalEvent; + + this.isImmediatePropagationStopped = returnTrue; + + if ( e && !this.isSimulated ) { + e.stopImmediatePropagation(); + } + + this.stopPropagation(); + } +}; + +// Includes all common event props including KeyEvent and MouseEvent specific props +jQuery.each( { + altKey: true, + bubbles: true, + cancelable: true, + changedTouches: true, + ctrlKey: true, + detail: true, + eventPhase: true, + metaKey: true, + pageX: true, + pageY: true, + shiftKey: true, + view: true, + "char": true, + charCode: true, + key: true, + keyCode: true, + button: true, + buttons: true, + clientX: true, + clientY: true, + offsetX: true, + offsetY: true, + pointerId: true, + pointerType: true, + screenX: true, + screenY: true, + targetTouches: true, + toElement: true, + touches: true, + + which: function( event ) { + var button = event.button; + + // Add which for key events + if ( event.which == null && rkeyEvent.test( event.type ) ) { + return event.charCode != null ? event.charCode : event.keyCode; + } + + // Add which for click: 1 === left; 2 === middle; 3 === right + if ( !event.which && button !== undefined && rmouseEvent.test( event.type ) ) { + if ( button & 1 ) { + return 1; + } + + if ( button & 2 ) { + return 3; + } + + if ( button & 4 ) { + return 2; + } + + return 0; + } + + return event.which; + } +}, jQuery.event.addProp ); + +// Create mouseenter/leave events using mouseover/out and event-time checks +// so that event delegation works in jQuery. +// Do the same for pointerenter/pointerleave and pointerover/pointerout +// +// Support: Safari 7 only +// Safari sends mouseenter too often; see: +// https://bugs.chromium.org/p/chromium/issues/detail?id=470258 +// for the description of the bug (it existed in older Chrome versions as well). +jQuery.each( { + mouseenter: "mouseover", + mouseleave: "mouseout", + pointerenter: "pointerover", + pointerleave: "pointerout" +}, function( orig, fix ) { + jQuery.event.special[ orig ] = { + delegateType: fix, + bindType: fix, + + handle: function( event ) { + var ret, + target = this, + related = event.relatedTarget, + handleObj = event.handleObj; + + // For mouseenter/leave call the handler if related is outside the target. + // NB: No relatedTarget if the mouse left/entered the browser window + if ( !related || ( related !== target && !jQuery.contains( target, related ) ) ) { + event.type = handleObj.origType; + ret = handleObj.handler.apply( this, arguments ); + event.type = fix; + } + return ret; + } + }; +} ); + +jQuery.fn.extend( { + + on: function( types, selector, data, fn ) { + return on( this, types, selector, data, fn ); + }, + one: function( types, selector, data, fn ) { + return on( this, types, selector, data, fn, 1 ); + }, + off: function( types, selector, fn ) { + var handleObj, type; + if ( types && types.preventDefault && types.handleObj ) { + + // ( event ) dispatched jQuery.Event + handleObj = types.handleObj; + jQuery( types.delegateTarget ).off( + handleObj.namespace ? + handleObj.origType + "." + handleObj.namespace : + handleObj.origType, + handleObj.selector, + handleObj.handler + ); + return this; + } + if ( typeof types === "object" ) { + + // ( types-object [, selector] ) + for ( type in types ) { + this.off( type, selector, types[ type ] ); + } + return this; + } + if ( selector === false || typeof selector === "function" ) { + + // ( types [, fn] ) + fn = selector; + selector = undefined; + } + if ( fn === false ) { + fn = returnFalse; + } + return this.each( function() { + jQuery.event.remove( this, types, fn, selector ); + } ); + } +} ); + + +var + + /* eslint-disable max-len */ + + // See https://github.com/eslint/eslint/issues/3229 + rxhtmlTag = /<(?!area|br|col|embed|hr|img|input|link|meta|param)(([a-z][^\/\0>\x20\t\r\n\f]*)[^>]*)\/>/gi, + + /* eslint-enable */ + + // Support: IE <=10 - 11, Edge 12 - 13 + // In IE/Edge using regex groups here causes severe slowdowns. + // See https://connect.microsoft.com/IE/feedback/details/1736512/ + rnoInnerhtml = /\s*$/g; + +// Prefer a tbody over its parent table for containing new rows +function manipulationTarget( elem, content ) { + if ( nodeName( elem, "table" ) && + nodeName( content.nodeType !== 11 ? content : content.firstChild, "tr" ) ) { + + return jQuery( ">tbody", elem )[ 0 ] || elem; + } + + return elem; +} + +// Replace/restore the type attribute of script elements for safe DOM manipulation +function disableScript( elem ) { + elem.type = ( elem.getAttribute( "type" ) !== null ) + "/" + elem.type; + return elem; +} +function restoreScript( elem ) { + var match = rscriptTypeMasked.exec( elem.type ); + + if ( match ) { + elem.type = match[ 1 ]; + } else { + elem.removeAttribute( "type" ); + } + + return elem; +} + +function cloneCopyEvent( src, dest ) { + var i, l, type, pdataOld, pdataCur, udataOld, udataCur, events; + + if ( dest.nodeType !== 1 ) { + return; + } + + // 1. Copy private data: events, handlers, etc. + if ( dataPriv.hasData( src ) ) { + pdataOld = dataPriv.access( src ); + pdataCur = dataPriv.set( dest, pdataOld ); + events = pdataOld.events; + + if ( events ) { + delete pdataCur.handle; + pdataCur.events = {}; + + for ( type in events ) { + for ( i = 0, l = events[ type ].length; i < l; i++ ) { + jQuery.event.add( dest, type, events[ type ][ i ] ); + } + } + } + } + + // 2. Copy user data + if ( dataUser.hasData( src ) ) { + udataOld = dataUser.access( src ); + udataCur = jQuery.extend( {}, udataOld ); + + dataUser.set( dest, udataCur ); + } +} + +// Fix IE bugs, see support tests +function fixInput( src, dest ) { + var nodeName = dest.nodeName.toLowerCase(); + + // Fails to persist the checked state of a cloned checkbox or radio button. + if ( nodeName === "input" && rcheckableType.test( src.type ) ) { + dest.checked = src.checked; + + // Fails to return the selected option to the default selected state when cloning options + } else if ( nodeName === "input" || nodeName === "textarea" ) { + dest.defaultValue = src.defaultValue; + } +} + +function domManip( collection, args, callback, ignored ) { + + // Flatten any nested arrays + args = concat.apply( [], args ); + + var fragment, first, scripts, hasScripts, node, doc, + i = 0, + l = collection.length, + iNoClone = l - 1, + value = args[ 0 ], + isFunction = jQuery.isFunction( value ); + + // We can't cloneNode fragments that contain checked, in WebKit + if ( isFunction || + ( l > 1 && typeof value === "string" && + !support.checkClone && rchecked.test( value ) ) ) { + return collection.each( function( index ) { + var self = collection.eq( index ); + if ( isFunction ) { + args[ 0 ] = value.call( this, index, self.html() ); + } + domManip( self, args, callback, ignored ); + } ); + } + + if ( l ) { + fragment = buildFragment( args, collection[ 0 ].ownerDocument, false, collection, ignored ); + first = fragment.firstChild; + + if ( fragment.childNodes.length === 1 ) { + fragment = first; + } + + // Require either new content or an interest in ignored elements to invoke the callback + if ( first || ignored ) { + scripts = jQuery.map( getAll( fragment, "script" ), disableScript ); + hasScripts = scripts.length; + + // Use the original fragment for the last item + // instead of the first because it can end up + // being emptied incorrectly in certain situations (#8070). + for ( ; i < l; i++ ) { + node = fragment; + + if ( i !== iNoClone ) { + node = jQuery.clone( node, true, true ); + + // Keep references to cloned scripts for later restoration + if ( hasScripts ) { + + // Support: Android <=4.0 only, PhantomJS 1 only + // push.apply(_, arraylike) throws on ancient WebKit + jQuery.merge( scripts, getAll( node, "script" ) ); + } + } + + callback.call( collection[ i ], node, i ); + } + + if ( hasScripts ) { + doc = scripts[ scripts.length - 1 ].ownerDocument; + + // Reenable scripts + jQuery.map( scripts, restoreScript ); + + // Evaluate executable scripts on first document insertion + for ( i = 0; i < hasScripts; i++ ) { + node = scripts[ i ]; + if ( rscriptType.test( node.type || "" ) && + !dataPriv.access( node, "globalEval" ) && + jQuery.contains( doc, node ) ) { + + if ( node.src ) { + + // Optional AJAX dependency, but won't run scripts if not present + if ( jQuery._evalUrl ) { + jQuery._evalUrl( node.src ); + } + } else { + DOMEval( node.textContent.replace( rcleanScript, "" ), doc ); + } + } + } + } + } + } + + return collection; +} + +function remove( elem, selector, keepData ) { + var node, + nodes = selector ? jQuery.filter( selector, elem ) : elem, + i = 0; + + for ( ; ( node = nodes[ i ] ) != null; i++ ) { + if ( !keepData && node.nodeType === 1 ) { + jQuery.cleanData( getAll( node ) ); + } + + if ( node.parentNode ) { + if ( keepData && jQuery.contains( node.ownerDocument, node ) ) { + setGlobalEval( getAll( node, "script" ) ); + } + node.parentNode.removeChild( node ); + } + } + + return elem; +} + +jQuery.extend( { + htmlPrefilter: function( html ) { + return html.replace( rxhtmlTag, "<$1>" ); + }, + + clone: function( elem, dataAndEvents, deepDataAndEvents ) { + var i, l, srcElements, destElements, + clone = elem.cloneNode( true ), + inPage = jQuery.contains( elem.ownerDocument, elem ); + + // Fix IE cloning issues + if ( !support.noCloneChecked && ( elem.nodeType === 1 || elem.nodeType === 11 ) && + !jQuery.isXMLDoc( elem ) ) { + + // We eschew Sizzle here for performance reasons: https://jsperf.com/getall-vs-sizzle/2 + destElements = getAll( clone ); + srcElements = getAll( elem ); + + for ( i = 0, l = srcElements.length; i < l; i++ ) { + fixInput( srcElements[ i ], destElements[ i ] ); + } + } + + // Copy the events from the original to the clone + if ( dataAndEvents ) { + if ( deepDataAndEvents ) { + srcElements = srcElements || getAll( elem ); + destElements = destElements || getAll( clone ); + + for ( i = 0, l = srcElements.length; i < l; i++ ) { + cloneCopyEvent( srcElements[ i ], destElements[ i ] ); + } + } else { + cloneCopyEvent( elem, clone ); + } + } + + // Preserve script evaluation history + destElements = getAll( clone, "script" ); + if ( destElements.length > 0 ) { + setGlobalEval( destElements, !inPage && getAll( elem, "script" ) ); + } + + // Return the cloned set + return clone; + }, + + cleanData: function( elems ) { + var data, elem, type, + special = jQuery.event.special, + i = 0; + + for ( ; ( elem = elems[ i ] ) !== undefined; i++ ) { + if ( acceptData( elem ) ) { + if ( ( data = elem[ dataPriv.expando ] ) ) { + if ( data.events ) { + for ( type in data.events ) { + if ( special[ type ] ) { + jQuery.event.remove( elem, type ); + + // This is a shortcut to avoid jQuery.event.remove's overhead + } else { + jQuery.removeEvent( elem, type, data.handle ); + } + } + } + + // Support: Chrome <=35 - 45+ + // Assign undefined instead of using delete, see Data#remove + elem[ dataPriv.expando ] = undefined; + } + if ( elem[ dataUser.expando ] ) { + + // Support: Chrome <=35 - 45+ + // Assign undefined instead of using delete, see Data#remove + elem[ dataUser.expando ] = undefined; + } + } + } + } +} ); + +jQuery.fn.extend( { + detach: function( selector ) { + return remove( this, selector, true ); + }, + + remove: function( selector ) { + return remove( this, selector ); + }, + + text: function( value ) { + return access( this, function( value ) { + return value === undefined ? + jQuery.text( this ) : + this.empty().each( function() { + if ( this.nodeType === 1 || this.nodeType === 11 || this.nodeType === 9 ) { + this.textContent = value; + } + } ); + }, null, value, arguments.length ); + }, + + append: function() { + return domManip( this, arguments, function( elem ) { + if ( this.nodeType === 1 || this.nodeType === 11 || this.nodeType === 9 ) { + var target = manipulationTarget( this, elem ); + target.appendChild( elem ); + } + } ); + }, + + prepend: function() { + return domManip( this, arguments, function( elem ) { + if ( this.nodeType === 1 || this.nodeType === 11 || this.nodeType === 9 ) { + var target = manipulationTarget( this, elem ); + target.insertBefore( elem, target.firstChild ); + } + } ); + }, + + before: function() { + return domManip( this, arguments, function( elem ) { + if ( this.parentNode ) { + this.parentNode.insertBefore( elem, this ); + } + } ); + }, + + after: function() { + return domManip( this, arguments, function( elem ) { + if ( this.parentNode ) { + this.parentNode.insertBefore( elem, this.nextSibling ); + } + } ); + }, + + empty: function() { + var elem, + i = 0; + + for ( ; ( elem = this[ i ] ) != null; i++ ) { + if ( elem.nodeType === 1 ) { + + // Prevent memory leaks + jQuery.cleanData( getAll( elem, false ) ); + + // Remove any remaining nodes + elem.textContent = ""; + } + } + + return this; + }, + + clone: function( dataAndEvents, deepDataAndEvents ) { + dataAndEvents = dataAndEvents == null ? false : dataAndEvents; + deepDataAndEvents = deepDataAndEvents == null ? dataAndEvents : deepDataAndEvents; + + return this.map( function() { + return jQuery.clone( this, dataAndEvents, deepDataAndEvents ); + } ); + }, + + html: function( value ) { + return access( this, function( value ) { + var elem = this[ 0 ] || {}, + i = 0, + l = this.length; + + if ( value === undefined && elem.nodeType === 1 ) { + return elem.innerHTML; + } + + // See if we can take a shortcut and just use innerHTML + if ( typeof value === "string" && !rnoInnerhtml.test( value ) && + !wrapMap[ ( rtagName.exec( value ) || [ "", "" ] )[ 1 ].toLowerCase() ] ) { + + value = jQuery.htmlPrefilter( value ); + + try { + for ( ; i < l; i++ ) { + elem = this[ i ] || {}; + + // Remove element nodes and prevent memory leaks + if ( elem.nodeType === 1 ) { + jQuery.cleanData( getAll( elem, false ) ); + elem.innerHTML = value; + } + } + + elem = 0; + + // If using innerHTML throws an exception, use the fallback method + } catch ( e ) {} + } + + if ( elem ) { + this.empty().append( value ); + } + }, null, value, arguments.length ); + }, + + replaceWith: function() { + var ignored = []; + + // Make the changes, replacing each non-ignored context element with the new content + return domManip( this, arguments, function( elem ) { + var parent = this.parentNode; + + if ( jQuery.inArray( this, ignored ) < 0 ) { + jQuery.cleanData( getAll( this ) ); + if ( parent ) { + parent.replaceChild( elem, this ); + } + } + + // Force callback invocation + }, ignored ); + } +} ); + +jQuery.each( { + appendTo: "append", + prependTo: "prepend", + insertBefore: "before", + insertAfter: "after", + replaceAll: "replaceWith" +}, function( name, original ) { + jQuery.fn[ name ] = function( selector ) { + var elems, + ret = [], + insert = jQuery( selector ), + last = insert.length - 1, + i = 0; + + for ( ; i <= last; i++ ) { + elems = i === last ? this : this.clone( true ); + jQuery( insert[ i ] )[ original ]( elems ); + + // Support: Android <=4.0 only, PhantomJS 1 only + // .get() because push.apply(_, arraylike) throws on ancient WebKit + push.apply( ret, elems.get() ); + } + + return this.pushStack( ret ); + }; +} ); +var rmargin = ( /^margin/ ); + +var rnumnonpx = new RegExp( "^(" + pnum + ")(?!px)[a-z%]+$", "i" ); + +var getStyles = function( elem ) { + + // Support: IE <=11 only, Firefox <=30 (#15098, #14150) + // IE throws on elements created in popups + // FF meanwhile throws on frame elements through "defaultView.getComputedStyle" + var view = elem.ownerDocument.defaultView; + + if ( !view || !view.opener ) { + view = window; + } + + return view.getComputedStyle( elem ); + }; + + + +( function() { + + // Executing both pixelPosition & boxSizingReliable tests require only one layout + // so they're executed at the same time to save the second computation. + function computeStyleTests() { + + // This is a singleton, we need to execute it only once + if ( !div ) { + return; + } + + div.style.cssText = + "box-sizing:border-box;" + + "position:relative;display:block;" + + "margin:auto;border:1px;padding:1px;" + + "top:1%;width:50%"; + div.innerHTML = ""; + documentElement.appendChild( container ); + + var divStyle = window.getComputedStyle( div ); + pixelPositionVal = divStyle.top !== "1%"; + + // Support: Android 4.0 - 4.3 only, Firefox <=3 - 44 + reliableMarginLeftVal = divStyle.marginLeft === "2px"; + boxSizingReliableVal = divStyle.width === "4px"; + + // Support: Android 4.0 - 4.3 only + // Some styles come back with percentage values, even though they shouldn't + div.style.marginRight = "50%"; + pixelMarginRightVal = divStyle.marginRight === "4px"; + + documentElement.removeChild( container ); + + // Nullify the div so it wouldn't be stored in the memory and + // it will also be a sign that checks already performed + div = null; + } + + var pixelPositionVal, boxSizingReliableVal, pixelMarginRightVal, reliableMarginLeftVal, + container = document.createElement( "div" ), + div = document.createElement( "div" ); + + // Finish early in limited (non-browser) environments + if ( !div.style ) { + return; + } + + // Support: IE <=9 - 11 only + // Style of cloned element affects source element cloned (#8908) + div.style.backgroundClip = "content-box"; + div.cloneNode( true ).style.backgroundClip = ""; + support.clearCloneStyle = div.style.backgroundClip === "content-box"; + + container.style.cssText = "border:0;width:8px;height:0;top:0;left:-9999px;" + + "padding:0;margin-top:1px;position:absolute"; + container.appendChild( div ); + + jQuery.extend( support, { + pixelPosition: function() { + computeStyleTests(); + return pixelPositionVal; + }, + boxSizingReliable: function() { + computeStyleTests(); + return boxSizingReliableVal; + }, + pixelMarginRight: function() { + computeStyleTests(); + return pixelMarginRightVal; + }, + reliableMarginLeft: function() { + computeStyleTests(); + return reliableMarginLeftVal; + } + } ); +} )(); + + +function curCSS( elem, name, computed ) { + var width, minWidth, maxWidth, ret, + + // Support: Firefox 51+ + // Retrieving style before computed somehow + // fixes an issue with getting wrong values + // on detached elements + style = elem.style; + + computed = computed || getStyles( elem ); + + // getPropertyValue is needed for: + // .css('filter') (IE 9 only, #12537) + // .css('--customProperty) (#3144) + if ( computed ) { + ret = computed.getPropertyValue( name ) || computed[ name ]; + + if ( ret === "" && !jQuery.contains( elem.ownerDocument, elem ) ) { + ret = jQuery.style( elem, name ); + } + + // A tribute to the "awesome hack by Dean Edwards" + // Android Browser returns percentage for some values, + // but width seems to be reliably pixels. + // This is against the CSSOM draft spec: + // https://drafts.csswg.org/cssom/#resolved-values + if ( !support.pixelMarginRight() && rnumnonpx.test( ret ) && rmargin.test( name ) ) { + + // Remember the original values + width = style.width; + minWidth = style.minWidth; + maxWidth = style.maxWidth; + + // Put in the new values to get a computed value out + style.minWidth = style.maxWidth = style.width = ret; + ret = computed.width; + + // Revert the changed values + style.width = width; + style.minWidth = minWidth; + style.maxWidth = maxWidth; + } + } + + return ret !== undefined ? + + // Support: IE <=9 - 11 only + // IE returns zIndex value as an integer. + ret + "" : + ret; +} + + +function addGetHookIf( conditionFn, hookFn ) { + + // Define the hook, we'll check on the first run if it's really needed. + return { + get: function() { + if ( conditionFn() ) { + + // Hook not needed (or it's not possible to use it due + // to missing dependency), remove it. + delete this.get; + return; + } + + // Hook needed; redefine it so that the support test is not executed again. + return ( this.get = hookFn ).apply( this, arguments ); + } + }; +} + + +var + + // Swappable if display is none or starts with table + // except "table", "table-cell", or "table-caption" + // See here for display values: https://developer.mozilla.org/en-US/docs/CSS/display + rdisplayswap = /^(none|table(?!-c[ea]).+)/, + rcustomProp = /^--/, + cssShow = { position: "absolute", visibility: "hidden", display: "block" }, + cssNormalTransform = { + letterSpacing: "0", + fontWeight: "400" + }, + + cssPrefixes = [ "Webkit", "Moz", "ms" ], + emptyStyle = document.createElement( "div" ).style; + +// Return a css property mapped to a potentially vendor prefixed property +function vendorPropName( name ) { + + // Shortcut for names that are not vendor prefixed + if ( name in emptyStyle ) { + return name; + } + + // Check for vendor prefixed names + var capName = name[ 0 ].toUpperCase() + name.slice( 1 ), + i = cssPrefixes.length; + + while ( i-- ) { + name = cssPrefixes[ i ] + capName; + if ( name in emptyStyle ) { + return name; + } + } +} + +// Return a property mapped along what jQuery.cssProps suggests or to +// a vendor prefixed property. +function finalPropName( name ) { + var ret = jQuery.cssProps[ name ]; + if ( !ret ) { + ret = jQuery.cssProps[ name ] = vendorPropName( name ) || name; + } + return ret; +} + +function setPositiveNumber( elem, value, subtract ) { + + // Any relative (+/-) values have already been + // normalized at this point + var matches = rcssNum.exec( value ); + return matches ? + + // Guard against undefined "subtract", e.g., when used as in cssHooks + Math.max( 0, matches[ 2 ] - ( subtract || 0 ) ) + ( matches[ 3 ] || "px" ) : + value; +} + +function augmentWidthOrHeight( elem, name, extra, isBorderBox, styles ) { + var i, + val = 0; + + // If we already have the right measurement, avoid augmentation + if ( extra === ( isBorderBox ? "border" : "content" ) ) { + i = 4; + + // Otherwise initialize for horizontal or vertical properties + } else { + i = name === "width" ? 1 : 0; + } + + for ( ; i < 4; i += 2 ) { + + // Both box models exclude margin, so add it if we want it + if ( extra === "margin" ) { + val += jQuery.css( elem, extra + cssExpand[ i ], true, styles ); + } + + if ( isBorderBox ) { + + // border-box includes padding, so remove it if we want content + if ( extra === "content" ) { + val -= jQuery.css( elem, "padding" + cssExpand[ i ], true, styles ); + } + + // At this point, extra isn't border nor margin, so remove border + if ( extra !== "margin" ) { + val -= jQuery.css( elem, "border" + cssExpand[ i ] + "Width", true, styles ); + } + } else { + + // At this point, extra isn't content, so add padding + val += jQuery.css( elem, "padding" + cssExpand[ i ], true, styles ); + + // At this point, extra isn't content nor padding, so add border + if ( extra !== "padding" ) { + val += jQuery.css( elem, "border" + cssExpand[ i ] + "Width", true, styles ); + } + } + } + + return val; +} + +function getWidthOrHeight( elem, name, extra ) { + + // Start with computed style + var valueIsBorderBox, + styles = getStyles( elem ), + val = curCSS( elem, name, styles ), + isBorderBox = jQuery.css( elem, "boxSizing", false, styles ) === "border-box"; + + // Computed unit is not pixels. Stop here and return. + if ( rnumnonpx.test( val ) ) { + return val; + } + + // Check for style in case a browser which returns unreliable values + // for getComputedStyle silently falls back to the reliable elem.style + valueIsBorderBox = isBorderBox && + ( support.boxSizingReliable() || val === elem.style[ name ] ); + + // Fall back to offsetWidth/Height when value is "auto" + // This happens for inline elements with no explicit setting (gh-3571) + if ( val === "auto" ) { + val = elem[ "offset" + name[ 0 ].toUpperCase() + name.slice( 1 ) ]; + } + + // Normalize "", auto, and prepare for extra + val = parseFloat( val ) || 0; + + // Use the active box-sizing model to add/subtract irrelevant styles + return ( val + + augmentWidthOrHeight( + elem, + name, + extra || ( isBorderBox ? "border" : "content" ), + valueIsBorderBox, + styles + ) + ) + "px"; +} + +jQuery.extend( { + + // Add in style property hooks for overriding the default + // behavior of getting and setting a style property + cssHooks: { + opacity: { + get: function( elem, computed ) { + if ( computed ) { + + // We should always get a number back from opacity + var ret = curCSS( elem, "opacity" ); + return ret === "" ? "1" : ret; + } + } + } + }, + + // Don't automatically add "px" to these possibly-unitless properties + cssNumber: { + "animationIterationCount": true, + "columnCount": true, + "fillOpacity": true, + "flexGrow": true, + "flexShrink": true, + "fontWeight": true, + "lineHeight": true, + "opacity": true, + "order": true, + "orphans": true, + "widows": true, + "zIndex": true, + "zoom": true + }, + + // Add in properties whose names you wish to fix before + // setting or getting the value + cssProps: { + "float": "cssFloat" + }, + + // Get and set the style property on a DOM Node + style: function( elem, name, value, extra ) { + + // Don't set styles on text and comment nodes + if ( !elem || elem.nodeType === 3 || elem.nodeType === 8 || !elem.style ) { + return; + } + + // Make sure that we're working with the right name + var ret, type, hooks, + origName = jQuery.camelCase( name ), + isCustomProp = rcustomProp.test( name ), + style = elem.style; + + // Make sure that we're working with the right name. We don't + // want to query the value if it is a CSS custom property + // since they are user-defined. + if ( !isCustomProp ) { + name = finalPropName( origName ); + } + + // Gets hook for the prefixed version, then unprefixed version + hooks = jQuery.cssHooks[ name ] || jQuery.cssHooks[ origName ]; + + // Check if we're setting a value + if ( value !== undefined ) { + type = typeof value; + + // Convert "+=" or "-=" to relative numbers (#7345) + if ( type === "string" && ( ret = rcssNum.exec( value ) ) && ret[ 1 ] ) { + value = adjustCSS( elem, name, ret ); + + // Fixes bug #9237 + type = "number"; + } + + // Make sure that null and NaN values aren't set (#7116) + if ( value == null || value !== value ) { + return; + } + + // If a number was passed in, add the unit (except for certain CSS properties) + if ( type === "number" ) { + value += ret && ret[ 3 ] || ( jQuery.cssNumber[ origName ] ? "" : "px" ); + } + + // background-* props affect original clone's values + if ( !support.clearCloneStyle && value === "" && name.indexOf( "background" ) === 0 ) { + style[ name ] = "inherit"; + } + + // If a hook was provided, use that value, otherwise just set the specified value + if ( !hooks || !( "set" in hooks ) || + ( value = hooks.set( elem, value, extra ) ) !== undefined ) { + + if ( isCustomProp ) { + style.setProperty( name, value ); + } else { + style[ name ] = value; + } + } + + } else { + + // If a hook was provided get the non-computed value from there + if ( hooks && "get" in hooks && + ( ret = hooks.get( elem, false, extra ) ) !== undefined ) { + + return ret; + } + + // Otherwise just get the value from the style object + return style[ name ]; + } + }, + + css: function( elem, name, extra, styles ) { + var val, num, hooks, + origName = jQuery.camelCase( name ), + isCustomProp = rcustomProp.test( name ); + + // Make sure that we're working with the right name. We don't + // want to modify the value if it is a CSS custom property + // since they are user-defined. + if ( !isCustomProp ) { + name = finalPropName( origName ); + } + + // Try prefixed name followed by the unprefixed name + hooks = jQuery.cssHooks[ name ] || jQuery.cssHooks[ origName ]; + + // If a hook was provided get the computed value from there + if ( hooks && "get" in hooks ) { + val = hooks.get( elem, true, extra ); + } + + // Otherwise, if a way to get the computed value exists, use that + if ( val === undefined ) { + val = curCSS( elem, name, styles ); + } + + // Convert "normal" to computed value + if ( val === "normal" && name in cssNormalTransform ) { + val = cssNormalTransform[ name ]; + } + + // Make numeric if forced or a qualifier was provided and val looks numeric + if ( extra === "" || extra ) { + num = parseFloat( val ); + return extra === true || isFinite( num ) ? num || 0 : val; + } + + return val; + } +} ); + +jQuery.each( [ "height", "width" ], function( i, name ) { + jQuery.cssHooks[ name ] = { + get: function( elem, computed, extra ) { + if ( computed ) { + + // Certain elements can have dimension info if we invisibly show them + // but it must have a current display style that would benefit + return rdisplayswap.test( jQuery.css( elem, "display" ) ) && + + // Support: Safari 8+ + // Table columns in Safari have non-zero offsetWidth & zero + // getBoundingClientRect().width unless display is changed. + // Support: IE <=11 only + // Running getBoundingClientRect on a disconnected node + // in IE throws an error. + ( !elem.getClientRects().length || !elem.getBoundingClientRect().width ) ? + swap( elem, cssShow, function() { + return getWidthOrHeight( elem, name, extra ); + } ) : + getWidthOrHeight( elem, name, extra ); + } + }, + + set: function( elem, value, extra ) { + var matches, + styles = extra && getStyles( elem ), + subtract = extra && augmentWidthOrHeight( + elem, + name, + extra, + jQuery.css( elem, "boxSizing", false, styles ) === "border-box", + styles + ); + + // Convert to pixels if value adjustment is needed + if ( subtract && ( matches = rcssNum.exec( value ) ) && + ( matches[ 3 ] || "px" ) !== "px" ) { + + elem.style[ name ] = value; + value = jQuery.css( elem, name ); + } + + return setPositiveNumber( elem, value, subtract ); + } + }; +} ); + +jQuery.cssHooks.marginLeft = addGetHookIf( support.reliableMarginLeft, + function( elem, computed ) { + if ( computed ) { + return ( parseFloat( curCSS( elem, "marginLeft" ) ) || + elem.getBoundingClientRect().left - + swap( elem, { marginLeft: 0 }, function() { + return elem.getBoundingClientRect().left; + } ) + ) + "px"; + } + } +); + +// These hooks are used by animate to expand properties +jQuery.each( { + margin: "", + padding: "", + border: "Width" +}, function( prefix, suffix ) { + jQuery.cssHooks[ prefix + suffix ] = { + expand: function( value ) { + var i = 0, + expanded = {}, + + // Assumes a single number if not a string + parts = typeof value === "string" ? value.split( " " ) : [ value ]; + + for ( ; i < 4; i++ ) { + expanded[ prefix + cssExpand[ i ] + suffix ] = + parts[ i ] || parts[ i - 2 ] || parts[ 0 ]; + } + + return expanded; + } + }; + + if ( !rmargin.test( prefix ) ) { + jQuery.cssHooks[ prefix + suffix ].set = setPositiveNumber; + } +} ); + +jQuery.fn.extend( { + css: function( name, value ) { + return access( this, function( elem, name, value ) { + var styles, len, + map = {}, + i = 0; + + if ( Array.isArray( name ) ) { + styles = getStyles( elem ); + len = name.length; + + for ( ; i < len; i++ ) { + map[ name[ i ] ] = jQuery.css( elem, name[ i ], false, styles ); + } + + return map; + } + + return value !== undefined ? + jQuery.style( elem, name, value ) : + jQuery.css( elem, name ); + }, name, value, arguments.length > 1 ); + } +} ); + + +function Tween( elem, options, prop, end, easing ) { + return new Tween.prototype.init( elem, options, prop, end, easing ); +} +jQuery.Tween = Tween; + +Tween.prototype = { + constructor: Tween, + init: function( elem, options, prop, end, easing, unit ) { + this.elem = elem; + this.prop = prop; + this.easing = easing || jQuery.easing._default; + this.options = options; + this.start = this.now = this.cur(); + this.end = end; + this.unit = unit || ( jQuery.cssNumber[ prop ] ? "" : "px" ); + }, + cur: function() { + var hooks = Tween.propHooks[ this.prop ]; + + return hooks && hooks.get ? + hooks.get( this ) : + Tween.propHooks._default.get( this ); + }, + run: function( percent ) { + var eased, + hooks = Tween.propHooks[ this.prop ]; + + if ( this.options.duration ) { + this.pos = eased = jQuery.easing[ this.easing ]( + percent, this.options.duration * percent, 0, 1, this.options.duration + ); + } else { + this.pos = eased = percent; + } + this.now = ( this.end - this.start ) * eased + this.start; + + if ( this.options.step ) { + this.options.step.call( this.elem, this.now, this ); + } + + if ( hooks && hooks.set ) { + hooks.set( this ); + } else { + Tween.propHooks._default.set( this ); + } + return this; + } +}; + +Tween.prototype.init.prototype = Tween.prototype; + +Tween.propHooks = { + _default: { + get: function( tween ) { + var result; + + // Use a property on the element directly when it is not a DOM element, + // or when there is no matching style property that exists. + if ( tween.elem.nodeType !== 1 || + tween.elem[ tween.prop ] != null && tween.elem.style[ tween.prop ] == null ) { + return tween.elem[ tween.prop ]; + } + + // Passing an empty string as a 3rd parameter to .css will automatically + // attempt a parseFloat and fallback to a string if the parse fails. + // Simple values such as "10px" are parsed to Float; + // complex values such as "rotate(1rad)" are returned as-is. + result = jQuery.css( tween.elem, tween.prop, "" ); + + // Empty strings, null, undefined and "auto" are converted to 0. + return !result || result === "auto" ? 0 : result; + }, + set: function( tween ) { + + // Use step hook for back compat. + // Use cssHook if its there. + // Use .style if available and use plain properties where available. + if ( jQuery.fx.step[ tween.prop ] ) { + jQuery.fx.step[ tween.prop ]( tween ); + } else if ( tween.elem.nodeType === 1 && + ( tween.elem.style[ jQuery.cssProps[ tween.prop ] ] != null || + jQuery.cssHooks[ tween.prop ] ) ) { + jQuery.style( tween.elem, tween.prop, tween.now + tween.unit ); + } else { + tween.elem[ tween.prop ] = tween.now; + } + } + } +}; + +// Support: IE <=9 only +// Panic based approach to setting things on disconnected nodes +Tween.propHooks.scrollTop = Tween.propHooks.scrollLeft = { + set: function( tween ) { + if ( tween.elem.nodeType && tween.elem.parentNode ) { + tween.elem[ tween.prop ] = tween.now; + } + } +}; + +jQuery.easing = { + linear: function( p ) { + return p; + }, + swing: function( p ) { + return 0.5 - Math.cos( p * Math.PI ) / 2; + }, + _default: "swing" +}; + +jQuery.fx = Tween.prototype.init; + +// Back compat <1.8 extension point +jQuery.fx.step = {}; + + + + +var + fxNow, inProgress, + rfxtypes = /^(?:toggle|show|hide)$/, + rrun = /queueHooks$/; + +function schedule() { + if ( inProgress ) { + if ( document.hidden === false && window.requestAnimationFrame ) { + window.requestAnimationFrame( schedule ); + } else { + window.setTimeout( schedule, jQuery.fx.interval ); + } + + jQuery.fx.tick(); + } +} + +// Animations created synchronously will run synchronously +function createFxNow() { + window.setTimeout( function() { + fxNow = undefined; + } ); + return ( fxNow = jQuery.now() ); +} + +// Generate parameters to create a standard animation +function genFx( type, includeWidth ) { + var which, + i = 0, + attrs = { height: type }; + + // If we include width, step value is 1 to do all cssExpand values, + // otherwise step value is 2 to skip over Left and Right + includeWidth = includeWidth ? 1 : 0; + for ( ; i < 4; i += 2 - includeWidth ) { + which = cssExpand[ i ]; + attrs[ "margin" + which ] = attrs[ "padding" + which ] = type; + } + + if ( includeWidth ) { + attrs.opacity = attrs.width = type; + } + + return attrs; +} + +function createTween( value, prop, animation ) { + var tween, + collection = ( Animation.tweeners[ prop ] || [] ).concat( Animation.tweeners[ "*" ] ), + index = 0, + length = collection.length; + for ( ; index < length; index++ ) { + if ( ( tween = collection[ index ].call( animation, prop, value ) ) ) { + + // We're done with this property + return tween; + } + } +} + +function defaultPrefilter( elem, props, opts ) { + var prop, value, toggle, hooks, oldfire, propTween, restoreDisplay, display, + isBox = "width" in props || "height" in props, + anim = this, + orig = {}, + style = elem.style, + hidden = elem.nodeType && isHiddenWithinTree( elem ), + dataShow = dataPriv.get( elem, "fxshow" ); + + // Queue-skipping animations hijack the fx hooks + if ( !opts.queue ) { + hooks = jQuery._queueHooks( elem, "fx" ); + if ( hooks.unqueued == null ) { + hooks.unqueued = 0; + oldfire = hooks.empty.fire; + hooks.empty.fire = function() { + if ( !hooks.unqueued ) { + oldfire(); + } + }; + } + hooks.unqueued++; + + anim.always( function() { + + // Ensure the complete handler is called before this completes + anim.always( function() { + hooks.unqueued--; + if ( !jQuery.queue( elem, "fx" ).length ) { + hooks.empty.fire(); + } + } ); + } ); + } + + // Detect show/hide animations + for ( prop in props ) { + value = props[ prop ]; + if ( rfxtypes.test( value ) ) { + delete props[ prop ]; + toggle = toggle || value === "toggle"; + if ( value === ( hidden ? "hide" : "show" ) ) { + + // Pretend to be hidden if this is a "show" and + // there is still data from a stopped show/hide + if ( value === "show" && dataShow && dataShow[ prop ] !== undefined ) { + hidden = true; + + // Ignore all other no-op show/hide data + } else { + continue; + } + } + orig[ prop ] = dataShow && dataShow[ prop ] || jQuery.style( elem, prop ); + } + } + + // Bail out if this is a no-op like .hide().hide() + propTween = !jQuery.isEmptyObject( props ); + if ( !propTween && jQuery.isEmptyObject( orig ) ) { + return; + } + + // Restrict "overflow" and "display" styles during box animations + if ( isBox && elem.nodeType === 1 ) { + + // Support: IE <=9 - 11, Edge 12 - 13 + // Record all 3 overflow attributes because IE does not infer the shorthand + // from identically-valued overflowX and overflowY + opts.overflow = [ style.overflow, style.overflowX, style.overflowY ]; + + // Identify a display type, preferring old show/hide data over the CSS cascade + restoreDisplay = dataShow && dataShow.display; + if ( restoreDisplay == null ) { + restoreDisplay = dataPriv.get( elem, "display" ); + } + display = jQuery.css( elem, "display" ); + if ( display === "none" ) { + if ( restoreDisplay ) { + display = restoreDisplay; + } else { + + // Get nonempty value(s) by temporarily forcing visibility + showHide( [ elem ], true ); + restoreDisplay = elem.style.display || restoreDisplay; + display = jQuery.css( elem, "display" ); + showHide( [ elem ] ); + } + } + + // Animate inline elements as inline-block + if ( display === "inline" || display === "inline-block" && restoreDisplay != null ) { + if ( jQuery.css( elem, "float" ) === "none" ) { + + // Restore the original display value at the end of pure show/hide animations + if ( !propTween ) { + anim.done( function() { + style.display = restoreDisplay; + } ); + if ( restoreDisplay == null ) { + display = style.display; + restoreDisplay = display === "none" ? "" : display; + } + } + style.display = "inline-block"; + } + } + } + + if ( opts.overflow ) { + style.overflow = "hidden"; + anim.always( function() { + style.overflow = opts.overflow[ 0 ]; + style.overflowX = opts.overflow[ 1 ]; + style.overflowY = opts.overflow[ 2 ]; + } ); + } + + // Implement show/hide animations + propTween = false; + for ( prop in orig ) { + + // General show/hide setup for this element animation + if ( !propTween ) { + if ( dataShow ) { + if ( "hidden" in dataShow ) { + hidden = dataShow.hidden; + } + } else { + dataShow = dataPriv.access( elem, "fxshow", { display: restoreDisplay } ); + } + + // Store hidden/visible for toggle so `.stop().toggle()` "reverses" + if ( toggle ) { + dataShow.hidden = !hidden; + } + + // Show elements before animating them + if ( hidden ) { + showHide( [ elem ], true ); + } + + /* eslint-disable no-loop-func */ + + anim.done( function() { + + /* eslint-enable no-loop-func */ + + // The final step of a "hide" animation is actually hiding the element + if ( !hidden ) { + showHide( [ elem ] ); + } + dataPriv.remove( elem, "fxshow" ); + for ( prop in orig ) { + jQuery.style( elem, prop, orig[ prop ] ); + } + } ); + } + + // Per-property setup + propTween = createTween( hidden ? dataShow[ prop ] : 0, prop, anim ); + if ( !( prop in dataShow ) ) { + dataShow[ prop ] = propTween.start; + if ( hidden ) { + propTween.end = propTween.start; + propTween.start = 0; + } + } + } +} + +function propFilter( props, specialEasing ) { + var index, name, easing, value, hooks; + + // camelCase, specialEasing and expand cssHook pass + for ( index in props ) { + name = jQuery.camelCase( index ); + easing = specialEasing[ name ]; + value = props[ index ]; + if ( Array.isArray( value ) ) { + easing = value[ 1 ]; + value = props[ index ] = value[ 0 ]; + } + + if ( index !== name ) { + props[ name ] = value; + delete props[ index ]; + } + + hooks = jQuery.cssHooks[ name ]; + if ( hooks && "expand" in hooks ) { + value = hooks.expand( value ); + delete props[ name ]; + + // Not quite $.extend, this won't overwrite existing keys. + // Reusing 'index' because we have the correct "name" + for ( index in value ) { + if ( !( index in props ) ) { + props[ index ] = value[ index ]; + specialEasing[ index ] = easing; + } + } + } else { + specialEasing[ name ] = easing; + } + } +} + +function Animation( elem, properties, options ) { + var result, + stopped, + index = 0, + length = Animation.prefilters.length, + deferred = jQuery.Deferred().always( function() { + + // Don't match elem in the :animated selector + delete tick.elem; + } ), + tick = function() { + if ( stopped ) { + return false; + } + var currentTime = fxNow || createFxNow(), + remaining = Math.max( 0, animation.startTime + animation.duration - currentTime ), + + // Support: Android 2.3 only + // Archaic crash bug won't allow us to use `1 - ( 0.5 || 0 )` (#12497) + temp = remaining / animation.duration || 0, + percent = 1 - temp, + index = 0, + length = animation.tweens.length; + + for ( ; index < length; index++ ) { + animation.tweens[ index ].run( percent ); + } + + deferred.notifyWith( elem, [ animation, percent, remaining ] ); + + // If there's more to do, yield + if ( percent < 1 && length ) { + return remaining; + } + + // If this was an empty animation, synthesize a final progress notification + if ( !length ) { + deferred.notifyWith( elem, [ animation, 1, 0 ] ); + } + + // Resolve the animation and report its conclusion + deferred.resolveWith( elem, [ animation ] ); + return false; + }, + animation = deferred.promise( { + elem: elem, + props: jQuery.extend( {}, properties ), + opts: jQuery.extend( true, { + specialEasing: {}, + easing: jQuery.easing._default + }, options ), + originalProperties: properties, + originalOptions: options, + startTime: fxNow || createFxNow(), + duration: options.duration, + tweens: [], + createTween: function( prop, end ) { + var tween = jQuery.Tween( elem, animation.opts, prop, end, + animation.opts.specialEasing[ prop ] || animation.opts.easing ); + animation.tweens.push( tween ); + return tween; + }, + stop: function( gotoEnd ) { + var index = 0, + + // If we are going to the end, we want to run all the tweens + // otherwise we skip this part + length = gotoEnd ? animation.tweens.length : 0; + if ( stopped ) { + return this; + } + stopped = true; + for ( ; index < length; index++ ) { + animation.tweens[ index ].run( 1 ); + } + + // Resolve when we played the last frame; otherwise, reject + if ( gotoEnd ) { + deferred.notifyWith( elem, [ animation, 1, 0 ] ); + deferred.resolveWith( elem, [ animation, gotoEnd ] ); + } else { + deferred.rejectWith( elem, [ animation, gotoEnd ] ); + } + return this; + } + } ), + props = animation.props; + + propFilter( props, animation.opts.specialEasing ); + + for ( ; index < length; index++ ) { + result = Animation.prefilters[ index ].call( animation, elem, props, animation.opts ); + if ( result ) { + if ( jQuery.isFunction( result.stop ) ) { + jQuery._queueHooks( animation.elem, animation.opts.queue ).stop = + jQuery.proxy( result.stop, result ); + } + return result; + } + } + + jQuery.map( props, createTween, animation ); + + if ( jQuery.isFunction( animation.opts.start ) ) { + animation.opts.start.call( elem, animation ); + } + + // Attach callbacks from options + animation + .progress( animation.opts.progress ) + .done( animation.opts.done, animation.opts.complete ) + .fail( animation.opts.fail ) + .always( animation.opts.always ); + + jQuery.fx.timer( + jQuery.extend( tick, { + elem: elem, + anim: animation, + queue: animation.opts.queue + } ) + ); + + return animation; +} + +jQuery.Animation = jQuery.extend( Animation, { + + tweeners: { + "*": [ function( prop, value ) { + var tween = this.createTween( prop, value ); + adjustCSS( tween.elem, prop, rcssNum.exec( value ), tween ); + return tween; + } ] + }, + + tweener: function( props, callback ) { + if ( jQuery.isFunction( props ) ) { + callback = props; + props = [ "*" ]; + } else { + props = props.match( rnothtmlwhite ); + } + + var prop, + index = 0, + length = props.length; + + for ( ; index < length; index++ ) { + prop = props[ index ]; + Animation.tweeners[ prop ] = Animation.tweeners[ prop ] || []; + Animation.tweeners[ prop ].unshift( callback ); + } + }, + + prefilters: [ defaultPrefilter ], + + prefilter: function( callback, prepend ) { + if ( prepend ) { + Animation.prefilters.unshift( callback ); + } else { + Animation.prefilters.push( callback ); + } + } +} ); + +jQuery.speed = function( speed, easing, fn ) { + var opt = speed && typeof speed === "object" ? jQuery.extend( {}, speed ) : { + complete: fn || !fn && easing || + jQuery.isFunction( speed ) && speed, + duration: speed, + easing: fn && easing || easing && !jQuery.isFunction( easing ) && easing + }; + + // Go to the end state if fx are off + if ( jQuery.fx.off ) { + opt.duration = 0; + + } else { + if ( typeof opt.duration !== "number" ) { + if ( opt.duration in jQuery.fx.speeds ) { + opt.duration = jQuery.fx.speeds[ opt.duration ]; + + } else { + opt.duration = jQuery.fx.speeds._default; + } + } + } + + // Normalize opt.queue - true/undefined/null -> "fx" + if ( opt.queue == null || opt.queue === true ) { + opt.queue = "fx"; + } + + // Queueing + opt.old = opt.complete; + + opt.complete = function() { + if ( jQuery.isFunction( opt.old ) ) { + opt.old.call( this ); + } + + if ( opt.queue ) { + jQuery.dequeue( this, opt.queue ); + } + }; + + return opt; +}; + +jQuery.fn.extend( { + fadeTo: function( speed, to, easing, callback ) { + + // Show any hidden elements after setting opacity to 0 + return this.filter( isHiddenWithinTree ).css( "opacity", 0 ).show() + + // Animate to the value specified + .end().animate( { opacity: to }, speed, easing, callback ); + }, + animate: function( prop, speed, easing, callback ) { + var empty = jQuery.isEmptyObject( prop ), + optall = jQuery.speed( speed, easing, callback ), + doAnimation = function() { + + // Operate on a copy of prop so per-property easing won't be lost + var anim = Animation( this, jQuery.extend( {}, prop ), optall ); + + // Empty animations, or finishing resolves immediately + if ( empty || dataPriv.get( this, "finish" ) ) { + anim.stop( true ); + } + }; + doAnimation.finish = doAnimation; + + return empty || optall.queue === false ? + this.each( doAnimation ) : + this.queue( optall.queue, doAnimation ); + }, + stop: function( type, clearQueue, gotoEnd ) { + var stopQueue = function( hooks ) { + var stop = hooks.stop; + delete hooks.stop; + stop( gotoEnd ); + }; + + if ( typeof type !== "string" ) { + gotoEnd = clearQueue; + clearQueue = type; + type = undefined; + } + if ( clearQueue && type !== false ) { + this.queue( type || "fx", [] ); + } + + return this.each( function() { + var dequeue = true, + index = type != null && type + "queueHooks", + timers = jQuery.timers, + data = dataPriv.get( this ); + + if ( index ) { + if ( data[ index ] && data[ index ].stop ) { + stopQueue( data[ index ] ); + } + } else { + for ( index in data ) { + if ( data[ index ] && data[ index ].stop && rrun.test( index ) ) { + stopQueue( data[ index ] ); + } + } + } + + for ( index = timers.length; index--; ) { + if ( timers[ index ].elem === this && + ( type == null || timers[ index ].queue === type ) ) { + + timers[ index ].anim.stop( gotoEnd ); + dequeue = false; + timers.splice( index, 1 ); + } + } + + // Start the next in the queue if the last step wasn't forced. + // Timers currently will call their complete callbacks, which + // will dequeue but only if they were gotoEnd. + if ( dequeue || !gotoEnd ) { + jQuery.dequeue( this, type ); + } + } ); + }, + finish: function( type ) { + if ( type !== false ) { + type = type || "fx"; + } + return this.each( function() { + var index, + data = dataPriv.get( this ), + queue = data[ type + "queue" ], + hooks = data[ type + "queueHooks" ], + timers = jQuery.timers, + length = queue ? queue.length : 0; + + // Enable finishing flag on private data + data.finish = true; + + // Empty the queue first + jQuery.queue( this, type, [] ); + + if ( hooks && hooks.stop ) { + hooks.stop.call( this, true ); + } + + // Look for any active animations, and finish them + for ( index = timers.length; index--; ) { + if ( timers[ index ].elem === this && timers[ index ].queue === type ) { + timers[ index ].anim.stop( true ); + timers.splice( index, 1 ); + } + } + + // Look for any animations in the old queue and finish them + for ( index = 0; index < length; index++ ) { + if ( queue[ index ] && queue[ index ].finish ) { + queue[ index ].finish.call( this ); + } + } + + // Turn off finishing flag + delete data.finish; + } ); + } +} ); + +jQuery.each( [ "toggle", "show", "hide" ], function( i, name ) { + var cssFn = jQuery.fn[ name ]; + jQuery.fn[ name ] = function( speed, easing, callback ) { + return speed == null || typeof speed === "boolean" ? + cssFn.apply( this, arguments ) : + this.animate( genFx( name, true ), speed, easing, callback ); + }; +} ); + +// Generate shortcuts for custom animations +jQuery.each( { + slideDown: genFx( "show" ), + slideUp: genFx( "hide" ), + slideToggle: genFx( "toggle" ), + fadeIn: { opacity: "show" }, + fadeOut: { opacity: "hide" }, + fadeToggle: { opacity: "toggle" } +}, function( name, props ) { + jQuery.fn[ name ] = function( speed, easing, callback ) { + return this.animate( props, speed, easing, callback ); + }; +} ); + +jQuery.timers = []; +jQuery.fx.tick = function() { + var timer, + i = 0, + timers = jQuery.timers; + + fxNow = jQuery.now(); + + for ( ; i < timers.length; i++ ) { + timer = timers[ i ]; + + // Run the timer and safely remove it when done (allowing for external removal) + if ( !timer() && timers[ i ] === timer ) { + timers.splice( i--, 1 ); + } + } + + if ( !timers.length ) { + jQuery.fx.stop(); + } + fxNow = undefined; +}; + +jQuery.fx.timer = function( timer ) { + jQuery.timers.push( timer ); + jQuery.fx.start(); +}; + +jQuery.fx.interval = 13; +jQuery.fx.start = function() { + if ( inProgress ) { + return; + } + + inProgress = true; + schedule(); +}; + +jQuery.fx.stop = function() { + inProgress = null; +}; + +jQuery.fx.speeds = { + slow: 600, + fast: 200, + + // Default speed + _default: 400 +}; + + +// Based off of the plugin by Clint Helfers, with permission. +// https://web.archive.org/web/20100324014747/http://blindsignals.com/index.php/2009/07/jquery-delay/ +jQuery.fn.delay = function( time, type ) { + time = jQuery.fx ? jQuery.fx.speeds[ time ] || time : time; + type = type || "fx"; + + return this.queue( type, function( next, hooks ) { + var timeout = window.setTimeout( next, time ); + hooks.stop = function() { + window.clearTimeout( timeout ); + }; + } ); +}; + + +( function() { + var input = document.createElement( "input" ), + select = document.createElement( "select" ), + opt = select.appendChild( document.createElement( "option" ) ); + + input.type = "checkbox"; + + // Support: Android <=4.3 only + // Default value for a checkbox should be "on" + support.checkOn = input.value !== ""; + + // Support: IE <=11 only + // Must access selectedIndex to make default options select + support.optSelected = opt.selected; + + // Support: IE <=11 only + // An input loses its value after becoming a radio + input = document.createElement( "input" ); + input.value = "t"; + input.type = "radio"; + support.radioValue = input.value === "t"; +} )(); + + +var boolHook, + attrHandle = jQuery.expr.attrHandle; + +jQuery.fn.extend( { + attr: function( name, value ) { + return access( this, jQuery.attr, name, value, arguments.length > 1 ); + }, + + removeAttr: function( name ) { + return this.each( function() { + jQuery.removeAttr( this, name ); + } ); + } +} ); + +jQuery.extend( { + attr: function( elem, name, value ) { + var ret, hooks, + nType = elem.nodeType; + + // Don't get/set attributes on text, comment and attribute nodes + if ( nType === 3 || nType === 8 || nType === 2 ) { + return; + } + + // Fallback to prop when attributes are not supported + if ( typeof elem.getAttribute === "undefined" ) { + return jQuery.prop( elem, name, value ); + } + + // Attribute hooks are determined by the lowercase version + // Grab necessary hook if one is defined + if ( nType !== 1 || !jQuery.isXMLDoc( elem ) ) { + hooks = jQuery.attrHooks[ name.toLowerCase() ] || + ( jQuery.expr.match.bool.test( name ) ? boolHook : undefined ); + } + + if ( value !== undefined ) { + if ( value === null ) { + jQuery.removeAttr( elem, name ); + return; + } + + if ( hooks && "set" in hooks && + ( ret = hooks.set( elem, value, name ) ) !== undefined ) { + return ret; + } + + elem.setAttribute( name, value + "" ); + return value; + } + + if ( hooks && "get" in hooks && ( ret = hooks.get( elem, name ) ) !== null ) { + return ret; + } + + ret = jQuery.find.attr( elem, name ); + + // Non-existent attributes return null, we normalize to undefined + return ret == null ? undefined : ret; + }, + + attrHooks: { + type: { + set: function( elem, value ) { + if ( !support.radioValue && value === "radio" && + nodeName( elem, "input" ) ) { + var val = elem.value; + elem.setAttribute( "type", value ); + if ( val ) { + elem.value = val; + } + return value; + } + } + } + }, + + removeAttr: function( elem, value ) { + var name, + i = 0, + + // Attribute names can contain non-HTML whitespace characters + // https://html.spec.whatwg.org/multipage/syntax.html#attributes-2 + attrNames = value && value.match( rnothtmlwhite ); + + if ( attrNames && elem.nodeType === 1 ) { + while ( ( name = attrNames[ i++ ] ) ) { + elem.removeAttribute( name ); + } + } + } +} ); + +// Hooks for boolean attributes +boolHook = { + set: function( elem, value, name ) { + if ( value === false ) { + + // Remove boolean attributes when set to false + jQuery.removeAttr( elem, name ); + } else { + elem.setAttribute( name, name ); + } + return name; + } +}; + +jQuery.each( jQuery.expr.match.bool.source.match( /\w+/g ), function( i, name ) { + var getter = attrHandle[ name ] || jQuery.find.attr; + + attrHandle[ name ] = function( elem, name, isXML ) { + var ret, handle, + lowercaseName = name.toLowerCase(); + + if ( !isXML ) { + + // Avoid an infinite loop by temporarily removing this function from the getter + handle = attrHandle[ lowercaseName ]; + attrHandle[ lowercaseName ] = ret; + ret = getter( elem, name, isXML ) != null ? + lowercaseName : + null; + attrHandle[ lowercaseName ] = handle; + } + return ret; + }; +} ); + + + + +var rfocusable = /^(?:input|select|textarea|button)$/i, + rclickable = /^(?:a|area)$/i; + +jQuery.fn.extend( { + prop: function( name, value ) { + return access( this, jQuery.prop, name, value, arguments.length > 1 ); + }, + + removeProp: function( name ) { + return this.each( function() { + delete this[ jQuery.propFix[ name ] || name ]; + } ); + } +} ); + +jQuery.extend( { + prop: function( elem, name, value ) { + var ret, hooks, + nType = elem.nodeType; + + // Don't get/set properties on text, comment and attribute nodes + if ( nType === 3 || nType === 8 || nType === 2 ) { + return; + } + + if ( nType !== 1 || !jQuery.isXMLDoc( elem ) ) { + + // Fix name and attach hooks + name = jQuery.propFix[ name ] || name; + hooks = jQuery.propHooks[ name ]; + } + + if ( value !== undefined ) { + if ( hooks && "set" in hooks && + ( ret = hooks.set( elem, value, name ) ) !== undefined ) { + return ret; + } + + return ( elem[ name ] = value ); + } + + if ( hooks && "get" in hooks && ( ret = hooks.get( elem, name ) ) !== null ) { + return ret; + } + + return elem[ name ]; + }, + + propHooks: { + tabIndex: { + get: function( elem ) { + + // Support: IE <=9 - 11 only + // elem.tabIndex doesn't always return the + // correct value when it hasn't been explicitly set + // https://web.archive.org/web/20141116233347/http://fluidproject.org/blog/2008/01/09/getting-setting-and-removing-tabindex-values-with-javascript/ + // Use proper attribute retrieval(#12072) + var tabindex = jQuery.find.attr( elem, "tabindex" ); + + if ( tabindex ) { + return parseInt( tabindex, 10 ); + } + + if ( + rfocusable.test( elem.nodeName ) || + rclickable.test( elem.nodeName ) && + elem.href + ) { + return 0; + } + + return -1; + } + } + }, + + propFix: { + "for": "htmlFor", + "class": "className" + } +} ); + +// Support: IE <=11 only +// Accessing the selectedIndex property +// forces the browser to respect setting selected +// on the option +// The getter ensures a default option is selected +// when in an optgroup +// eslint rule "no-unused-expressions" is disabled for this code +// since it considers such accessions noop +if ( !support.optSelected ) { + jQuery.propHooks.selected = { + get: function( elem ) { + + /* eslint no-unused-expressions: "off" */ + + var parent = elem.parentNode; + if ( parent && parent.parentNode ) { + parent.parentNode.selectedIndex; + } + return null; + }, + set: function( elem ) { + + /* eslint no-unused-expressions: "off" */ + + var parent = elem.parentNode; + if ( parent ) { + parent.selectedIndex; + + if ( parent.parentNode ) { + parent.parentNode.selectedIndex; + } + } + } + }; +} + +jQuery.each( [ + "tabIndex", + "readOnly", + "maxLength", + "cellSpacing", + "cellPadding", + "rowSpan", + "colSpan", + "useMap", + "frameBorder", + "contentEditable" +], function() { + jQuery.propFix[ this.toLowerCase() ] = this; +} ); + + + + + // Strip and collapse whitespace according to HTML spec + // https://html.spec.whatwg.org/multipage/infrastructure.html#strip-and-collapse-whitespace + function stripAndCollapse( value ) { + var tokens = value.match( rnothtmlwhite ) || []; + return tokens.join( " " ); + } + + +function getClass( elem ) { + return elem.getAttribute && elem.getAttribute( "class" ) || ""; +} + +jQuery.fn.extend( { + addClass: function( value ) { + var classes, elem, cur, curValue, clazz, j, finalValue, + i = 0; + + if ( jQuery.isFunction( value ) ) { + return this.each( function( j ) { + jQuery( this ).addClass( value.call( this, j, getClass( this ) ) ); + } ); + } + + if ( typeof value === "string" && value ) { + classes = value.match( rnothtmlwhite ) || []; + + while ( ( elem = this[ i++ ] ) ) { + curValue = getClass( elem ); + cur = elem.nodeType === 1 && ( " " + stripAndCollapse( curValue ) + " " ); + + if ( cur ) { + j = 0; + while ( ( clazz = classes[ j++ ] ) ) { + if ( cur.indexOf( " " + clazz + " " ) < 0 ) { + cur += clazz + " "; + } + } + + // Only assign if different to avoid unneeded rendering. + finalValue = stripAndCollapse( cur ); + if ( curValue !== finalValue ) { + elem.setAttribute( "class", finalValue ); + } + } + } + } + + return this; + }, + + removeClass: function( value ) { + var classes, elem, cur, curValue, clazz, j, finalValue, + i = 0; + + if ( jQuery.isFunction( value ) ) { + return this.each( function( j ) { + jQuery( this ).removeClass( value.call( this, j, getClass( this ) ) ); + } ); + } + + if ( !arguments.length ) { + return this.attr( "class", "" ); + } + + if ( typeof value === "string" && value ) { + classes = value.match( rnothtmlwhite ) || []; + + while ( ( elem = this[ i++ ] ) ) { + curValue = getClass( elem ); + + // This expression is here for better compressibility (see addClass) + cur = elem.nodeType === 1 && ( " " + stripAndCollapse( curValue ) + " " ); + + if ( cur ) { + j = 0; + while ( ( clazz = classes[ j++ ] ) ) { + + // Remove *all* instances + while ( cur.indexOf( " " + clazz + " " ) > -1 ) { + cur = cur.replace( " " + clazz + " ", " " ); + } + } + + // Only assign if different to avoid unneeded rendering. + finalValue = stripAndCollapse( cur ); + if ( curValue !== finalValue ) { + elem.setAttribute( "class", finalValue ); + } + } + } + } + + return this; + }, + + toggleClass: function( value, stateVal ) { + var type = typeof value; + + if ( typeof stateVal === "boolean" && type === "string" ) { + return stateVal ? this.addClass( value ) : this.removeClass( value ); + } + + if ( jQuery.isFunction( value ) ) { + return this.each( function( i ) { + jQuery( this ).toggleClass( + value.call( this, i, getClass( this ), stateVal ), + stateVal + ); + } ); + } + + return this.each( function() { + var className, i, self, classNames; + + if ( type === "string" ) { + + // Toggle individual class names + i = 0; + self = jQuery( this ); + classNames = value.match( rnothtmlwhite ) || []; + + while ( ( className = classNames[ i++ ] ) ) { + + // Check each className given, space separated list + if ( self.hasClass( className ) ) { + self.removeClass( className ); + } else { + self.addClass( className ); + } + } + + // Toggle whole class name + } else if ( value === undefined || type === "boolean" ) { + className = getClass( this ); + if ( className ) { + + // Store className if set + dataPriv.set( this, "__className__", className ); + } + + // If the element has a class name or if we're passed `false`, + // then remove the whole classname (if there was one, the above saved it). + // Otherwise bring back whatever was previously saved (if anything), + // falling back to the empty string if nothing was stored. + if ( this.setAttribute ) { + this.setAttribute( "class", + className || value === false ? + "" : + dataPriv.get( this, "__className__" ) || "" + ); + } + } + } ); + }, + + hasClass: function( selector ) { + var className, elem, + i = 0; + + className = " " + selector + " "; + while ( ( elem = this[ i++ ] ) ) { + if ( elem.nodeType === 1 && + ( " " + stripAndCollapse( getClass( elem ) ) + " " ).indexOf( className ) > -1 ) { + return true; + } + } + + return false; + } +} ); + + + + +var rreturn = /\r/g; + +jQuery.fn.extend( { + val: function( value ) { + var hooks, ret, isFunction, + elem = this[ 0 ]; + + if ( !arguments.length ) { + if ( elem ) { + hooks = jQuery.valHooks[ elem.type ] || + jQuery.valHooks[ elem.nodeName.toLowerCase() ]; + + if ( hooks && + "get" in hooks && + ( ret = hooks.get( elem, "value" ) ) !== undefined + ) { + return ret; + } + + ret = elem.value; + + // Handle most common string cases + if ( typeof ret === "string" ) { + return ret.replace( rreturn, "" ); + } + + // Handle cases where value is null/undef or number + return ret == null ? "" : ret; + } + + return; + } + + isFunction = jQuery.isFunction( value ); + + return this.each( function( i ) { + var val; + + if ( this.nodeType !== 1 ) { + return; + } + + if ( isFunction ) { + val = value.call( this, i, jQuery( this ).val() ); + } else { + val = value; + } + + // Treat null/undefined as ""; convert numbers to string + if ( val == null ) { + val = ""; + + } else if ( typeof val === "number" ) { + val += ""; + + } else if ( Array.isArray( val ) ) { + val = jQuery.map( val, function( value ) { + return value == null ? "" : value + ""; + } ); + } + + hooks = jQuery.valHooks[ this.type ] || jQuery.valHooks[ this.nodeName.toLowerCase() ]; + + // If set returns undefined, fall back to normal setting + if ( !hooks || !( "set" in hooks ) || hooks.set( this, val, "value" ) === undefined ) { + this.value = val; + } + } ); + } +} ); + +jQuery.extend( { + valHooks: { + option: { + get: function( elem ) { + + var val = jQuery.find.attr( elem, "value" ); + return val != null ? + val : + + // Support: IE <=10 - 11 only + // option.text throws exceptions (#14686, #14858) + // Strip and collapse whitespace + // https://html.spec.whatwg.org/#strip-and-collapse-whitespace + stripAndCollapse( jQuery.text( elem ) ); + } + }, + select: { + get: function( elem ) { + var value, option, i, + options = elem.options, + index = elem.selectedIndex, + one = elem.type === "select-one", + values = one ? null : [], + max = one ? index + 1 : options.length; + + if ( index < 0 ) { + i = max; + + } else { + i = one ? index : 0; + } + + // Loop through all the selected options + for ( ; i < max; i++ ) { + option = options[ i ]; + + // Support: IE <=9 only + // IE8-9 doesn't update selected after form reset (#2551) + if ( ( option.selected || i === index ) && + + // Don't return options that are disabled or in a disabled optgroup + !option.disabled && + ( !option.parentNode.disabled || + !nodeName( option.parentNode, "optgroup" ) ) ) { + + // Get the specific value for the option + value = jQuery( option ).val(); + + // We don't need an array for one selects + if ( one ) { + return value; + } + + // Multi-Selects return an array + values.push( value ); + } + } + + return values; + }, + + set: function( elem, value ) { + var optionSet, option, + options = elem.options, + values = jQuery.makeArray( value ), + i = options.length; + + while ( i-- ) { + option = options[ i ]; + + /* eslint-disable no-cond-assign */ + + if ( option.selected = + jQuery.inArray( jQuery.valHooks.option.get( option ), values ) > -1 + ) { + optionSet = true; + } + + /* eslint-enable no-cond-assign */ + } + + // Force browsers to behave consistently when non-matching value is set + if ( !optionSet ) { + elem.selectedIndex = -1; + } + return values; + } + } + } +} ); + +// Radios and checkboxes getter/setter +jQuery.each( [ "radio", "checkbox" ], function() { + jQuery.valHooks[ this ] = { + set: function( elem, value ) { + if ( Array.isArray( value ) ) { + return ( elem.checked = jQuery.inArray( jQuery( elem ).val(), value ) > -1 ); + } + } + }; + if ( !support.checkOn ) { + jQuery.valHooks[ this ].get = function( elem ) { + return elem.getAttribute( "value" ) === null ? "on" : elem.value; + }; + } +} ); + + + + +// Return jQuery for attributes-only inclusion + + +var rfocusMorph = /^(?:focusinfocus|focusoutblur)$/; + +jQuery.extend( jQuery.event, { + + trigger: function( event, data, elem, onlyHandlers ) { + + var i, cur, tmp, bubbleType, ontype, handle, special, + eventPath = [ elem || document ], + type = hasOwn.call( event, "type" ) ? event.type : event, + namespaces = hasOwn.call( event, "namespace" ) ? event.namespace.split( "." ) : []; + + cur = tmp = elem = elem || document; + + // Don't do events on text and comment nodes + if ( elem.nodeType === 3 || elem.nodeType === 8 ) { + return; + } + + // focus/blur morphs to focusin/out; ensure we're not firing them right now + if ( rfocusMorph.test( type + jQuery.event.triggered ) ) { + return; + } + + if ( type.indexOf( "." ) > -1 ) { + + // Namespaced trigger; create a regexp to match event type in handle() + namespaces = type.split( "." ); + type = namespaces.shift(); + namespaces.sort(); + } + ontype = type.indexOf( ":" ) < 0 && "on" + type; + + // Caller can pass in a jQuery.Event object, Object, or just an event type string + event = event[ jQuery.expando ] ? + event : + new jQuery.Event( type, typeof event === "object" && event ); + + // Trigger bitmask: & 1 for native handlers; & 2 for jQuery (always true) + event.isTrigger = onlyHandlers ? 2 : 3; + event.namespace = namespaces.join( "." ); + event.rnamespace = event.namespace ? + new RegExp( "(^|\\.)" + namespaces.join( "\\.(?:.*\\.|)" ) + "(\\.|$)" ) : + null; + + // Clean up the event in case it is being reused + event.result = undefined; + if ( !event.target ) { + event.target = elem; + } + + // Clone any incoming data and prepend the event, creating the handler arg list + data = data == null ? + [ event ] : + jQuery.makeArray( data, [ event ] ); + + // Allow special events to draw outside the lines + special = jQuery.event.special[ type ] || {}; + if ( !onlyHandlers && special.trigger && special.trigger.apply( elem, data ) === false ) { + return; + } + + // Determine event propagation path in advance, per W3C events spec (#9951) + // Bubble up to document, then to window; watch for a global ownerDocument var (#9724) + if ( !onlyHandlers && !special.noBubble && !jQuery.isWindow( elem ) ) { + + bubbleType = special.delegateType || type; + if ( !rfocusMorph.test( bubbleType + type ) ) { + cur = cur.parentNode; + } + for ( ; cur; cur = cur.parentNode ) { + eventPath.push( cur ); + tmp = cur; + } + + // Only add window if we got to document (e.g., not plain obj or detached DOM) + if ( tmp === ( elem.ownerDocument || document ) ) { + eventPath.push( tmp.defaultView || tmp.parentWindow || window ); + } + } + + // Fire handlers on the event path + i = 0; + while ( ( cur = eventPath[ i++ ] ) && !event.isPropagationStopped() ) { + + event.type = i > 1 ? + bubbleType : + special.bindType || type; + + // jQuery handler + handle = ( dataPriv.get( cur, "events" ) || {} )[ event.type ] && + dataPriv.get( cur, "handle" ); + if ( handle ) { + handle.apply( cur, data ); + } + + // Native handler + handle = ontype && cur[ ontype ]; + if ( handle && handle.apply && acceptData( cur ) ) { + event.result = handle.apply( cur, data ); + if ( event.result === false ) { + event.preventDefault(); + } + } + } + event.type = type; + + // If nobody prevented the default action, do it now + if ( !onlyHandlers && !event.isDefaultPrevented() ) { + + if ( ( !special._default || + special._default.apply( eventPath.pop(), data ) === false ) && + acceptData( elem ) ) { + + // Call a native DOM method on the target with the same name as the event. + // Don't do default actions on window, that's where global variables be (#6170) + if ( ontype && jQuery.isFunction( elem[ type ] ) && !jQuery.isWindow( elem ) ) { + + // Don't re-trigger an onFOO event when we call its FOO() method + tmp = elem[ ontype ]; + + if ( tmp ) { + elem[ ontype ] = null; + } + + // Prevent re-triggering of the same event, since we already bubbled it above + jQuery.event.triggered = type; + elem[ type ](); + jQuery.event.triggered = undefined; + + if ( tmp ) { + elem[ ontype ] = tmp; + } + } + } + } + + return event.result; + }, + + // Piggyback on a donor event to simulate a different one + // Used only for `focus(in | out)` events + simulate: function( type, elem, event ) { + var e = jQuery.extend( + new jQuery.Event(), + event, + { + type: type, + isSimulated: true + } + ); + + jQuery.event.trigger( e, null, elem ); + } + +} ); + +jQuery.fn.extend( { + + trigger: function( type, data ) { + return this.each( function() { + jQuery.event.trigger( type, data, this ); + } ); + }, + triggerHandler: function( type, data ) { + var elem = this[ 0 ]; + if ( elem ) { + return jQuery.event.trigger( type, data, elem, true ); + } + } +} ); + + +jQuery.each( ( "blur focus focusin focusout resize scroll click dblclick " + + "mousedown mouseup mousemove mouseover mouseout mouseenter mouseleave " + + "change select submit keydown keypress keyup contextmenu" ).split( " " ), + function( i, name ) { + + // Handle event binding + jQuery.fn[ name ] = function( data, fn ) { + return arguments.length > 0 ? + this.on( name, null, data, fn ) : + this.trigger( name ); + }; +} ); + +jQuery.fn.extend( { + hover: function( fnOver, fnOut ) { + return this.mouseenter( fnOver ).mouseleave( fnOut || fnOver ); + } +} ); + + + + +support.focusin = "onfocusin" in window; + + +// Support: Firefox <=44 +// Firefox doesn't have focus(in | out) events +// Related ticket - https://bugzilla.mozilla.org/show_bug.cgi?id=687787 +// +// Support: Chrome <=48 - 49, Safari <=9.0 - 9.1 +// focus(in | out) events fire after focus & blur events, +// which is spec violation - http://www.w3.org/TR/DOM-Level-3-Events/#events-focusevent-event-order +// Related ticket - https://bugs.chromium.org/p/chromium/issues/detail?id=449857 +if ( !support.focusin ) { + jQuery.each( { focus: "focusin", blur: "focusout" }, function( orig, fix ) { + + // Attach a single capturing handler on the document while someone wants focusin/focusout + var handler = function( event ) { + jQuery.event.simulate( fix, event.target, jQuery.event.fix( event ) ); + }; + + jQuery.event.special[ fix ] = { + setup: function() { + var doc = this.ownerDocument || this, + attaches = dataPriv.access( doc, fix ); + + if ( !attaches ) { + doc.addEventListener( orig, handler, true ); + } + dataPriv.access( doc, fix, ( attaches || 0 ) + 1 ); + }, + teardown: function() { + var doc = this.ownerDocument || this, + attaches = dataPriv.access( doc, fix ) - 1; + + if ( !attaches ) { + doc.removeEventListener( orig, handler, true ); + dataPriv.remove( doc, fix ); + + } else { + dataPriv.access( doc, fix, attaches ); + } + } + }; + } ); +} +var location = window.location; + +var nonce = jQuery.now(); + +var rquery = ( /\?/ ); + + + +// Cross-browser xml parsing +jQuery.parseXML = function( data ) { + var xml; + if ( !data || typeof data !== "string" ) { + return null; + } + + // Support: IE 9 - 11 only + // IE throws on parseFromString with invalid input. + try { + xml = ( new window.DOMParser() ).parseFromString( data, "text/xml" ); + } catch ( e ) { + xml = undefined; + } + + if ( !xml || xml.getElementsByTagName( "parsererror" ).length ) { + jQuery.error( "Invalid XML: " + data ); + } + return xml; +}; + + +var + rbracket = /\[\]$/, + rCRLF = /\r?\n/g, + rsubmitterTypes = /^(?:submit|button|image|reset|file)$/i, + rsubmittable = /^(?:input|select|textarea|keygen)/i; + +function buildParams( prefix, obj, traditional, add ) { + var name; + + if ( Array.isArray( obj ) ) { + + // Serialize array item. + jQuery.each( obj, function( i, v ) { + if ( traditional || rbracket.test( prefix ) ) { + + // Treat each array item as a scalar. + add( prefix, v ); + + } else { + + // Item is non-scalar (array or object), encode its numeric index. + buildParams( + prefix + "[" + ( typeof v === "object" && v != null ? i : "" ) + "]", + v, + traditional, + add + ); + } + } ); + + } else if ( !traditional && jQuery.type( obj ) === "object" ) { + + // Serialize object item. + for ( name in obj ) { + buildParams( prefix + "[" + name + "]", obj[ name ], traditional, add ); + } + + } else { + + // Serialize scalar item. + add( prefix, obj ); + } +} + +// Serialize an array of form elements or a set of +// key/values into a query string +jQuery.param = function( a, traditional ) { + var prefix, + s = [], + add = function( key, valueOrFunction ) { + + // If value is a function, invoke it and use its return value + var value = jQuery.isFunction( valueOrFunction ) ? + valueOrFunction() : + valueOrFunction; + + s[ s.length ] = encodeURIComponent( key ) + "=" + + encodeURIComponent( value == null ? "" : value ); + }; + + // If an array was passed in, assume that it is an array of form elements. + if ( Array.isArray( a ) || ( a.jquery && !jQuery.isPlainObject( a ) ) ) { + + // Serialize the form elements + jQuery.each( a, function() { + add( this.name, this.value ); + } ); + + } else { + + // If traditional, encode the "old" way (the way 1.3.2 or older + // did it), otherwise encode params recursively. + for ( prefix in a ) { + buildParams( prefix, a[ prefix ], traditional, add ); + } + } + + // Return the resulting serialization + return s.join( "&" ); +}; + +jQuery.fn.extend( { + serialize: function() { + return jQuery.param( this.serializeArray() ); + }, + serializeArray: function() { + return this.map( function() { + + // Can add propHook for "elements" to filter or add form elements + var elements = jQuery.prop( this, "elements" ); + return elements ? jQuery.makeArray( elements ) : this; + } ) + .filter( function() { + var type = this.type; + + // Use .is( ":disabled" ) so that fieldset[disabled] works + return this.name && !jQuery( this ).is( ":disabled" ) && + rsubmittable.test( this.nodeName ) && !rsubmitterTypes.test( type ) && + ( this.checked || !rcheckableType.test( type ) ); + } ) + .map( function( i, elem ) { + var val = jQuery( this ).val(); + + if ( val == null ) { + return null; + } + + if ( Array.isArray( val ) ) { + return jQuery.map( val, function( val ) { + return { name: elem.name, value: val.replace( rCRLF, "\r\n" ) }; + } ); + } + + return { name: elem.name, value: val.replace( rCRLF, "\r\n" ) }; + } ).get(); + } +} ); + + +var + r20 = /%20/g, + rhash = /#.*$/, + rantiCache = /([?&])_=[^&]*/, + rheaders = /^(.*?):[ \t]*([^\r\n]*)$/mg, + + // #7653, #8125, #8152: local protocol detection + rlocalProtocol = /^(?:about|app|app-storage|.+-extension|file|res|widget):$/, + rnoContent = /^(?:GET|HEAD)$/, + rprotocol = /^\/\//, + + /* Prefilters + * 1) They are useful to introduce custom dataTypes (see ajax/jsonp.js for an example) + * 2) These are called: + * - BEFORE asking for a transport + * - AFTER param serialization (s.data is a string if s.processData is true) + * 3) key is the dataType + * 4) the catchall symbol "*" can be used + * 5) execution will start with transport dataType and THEN continue down to "*" if needed + */ + prefilters = {}, + + /* Transports bindings + * 1) key is the dataType + * 2) the catchall symbol "*" can be used + * 3) selection will start with transport dataType and THEN go to "*" if needed + */ + transports = {}, + + // Avoid comment-prolog char sequence (#10098); must appease lint and evade compression + allTypes = "*/".concat( "*" ), + + // Anchor tag for parsing the document origin + originAnchor = document.createElement( "a" ); + originAnchor.href = location.href; + +// Base "constructor" for jQuery.ajaxPrefilter and jQuery.ajaxTransport +function addToPrefiltersOrTransports( structure ) { + + // dataTypeExpression is optional and defaults to "*" + return function( dataTypeExpression, func ) { + + if ( typeof dataTypeExpression !== "string" ) { + func = dataTypeExpression; + dataTypeExpression = "*"; + } + + var dataType, + i = 0, + dataTypes = dataTypeExpression.toLowerCase().match( rnothtmlwhite ) || []; + + if ( jQuery.isFunction( func ) ) { + + // For each dataType in the dataTypeExpression + while ( ( dataType = dataTypes[ i++ ] ) ) { + + // Prepend if requested + if ( dataType[ 0 ] === "+" ) { + dataType = dataType.slice( 1 ) || "*"; + ( structure[ dataType ] = structure[ dataType ] || [] ).unshift( func ); + + // Otherwise append + } else { + ( structure[ dataType ] = structure[ dataType ] || [] ).push( func ); + } + } + } + }; +} + +// Base inspection function for prefilters and transports +function inspectPrefiltersOrTransports( structure, options, originalOptions, jqXHR ) { + + var inspected = {}, + seekingTransport = ( structure === transports ); + + function inspect( dataType ) { + var selected; + inspected[ dataType ] = true; + jQuery.each( structure[ dataType ] || [], function( _, prefilterOrFactory ) { + var dataTypeOrTransport = prefilterOrFactory( options, originalOptions, jqXHR ); + if ( typeof dataTypeOrTransport === "string" && + !seekingTransport && !inspected[ dataTypeOrTransport ] ) { + + options.dataTypes.unshift( dataTypeOrTransport ); + inspect( dataTypeOrTransport ); + return false; + } else if ( seekingTransport ) { + return !( selected = dataTypeOrTransport ); + } + } ); + return selected; + } + + return inspect( options.dataTypes[ 0 ] ) || !inspected[ "*" ] && inspect( "*" ); +} + +// A special extend for ajax options +// that takes "flat" options (not to be deep extended) +// Fixes #9887 +function ajaxExtend( target, src ) { + var key, deep, + flatOptions = jQuery.ajaxSettings.flatOptions || {}; + + for ( key in src ) { + if ( src[ key ] !== undefined ) { + ( flatOptions[ key ] ? target : ( deep || ( deep = {} ) ) )[ key ] = src[ key ]; + } + } + if ( deep ) { + jQuery.extend( true, target, deep ); + } + + return target; +} + +/* Handles responses to an ajax request: + * - finds the right dataType (mediates between content-type and expected dataType) + * - returns the corresponding response + */ +function ajaxHandleResponses( s, jqXHR, responses ) { + + var ct, type, finalDataType, firstDataType, + contents = s.contents, + dataTypes = s.dataTypes; + + // Remove auto dataType and get content-type in the process + while ( dataTypes[ 0 ] === "*" ) { + dataTypes.shift(); + if ( ct === undefined ) { + ct = s.mimeType || jqXHR.getResponseHeader( "Content-Type" ); + } + } + + // Check if we're dealing with a known content-type + if ( ct ) { + for ( type in contents ) { + if ( contents[ type ] && contents[ type ].test( ct ) ) { + dataTypes.unshift( type ); + break; + } + } + } + + // Check to see if we have a response for the expected dataType + if ( dataTypes[ 0 ] in responses ) { + finalDataType = dataTypes[ 0 ]; + } else { + + // Try convertible dataTypes + for ( type in responses ) { + if ( !dataTypes[ 0 ] || s.converters[ type + " " + dataTypes[ 0 ] ] ) { + finalDataType = type; + break; + } + if ( !firstDataType ) { + firstDataType = type; + } + } + + // Or just use first one + finalDataType = finalDataType || firstDataType; + } + + // If we found a dataType + // We add the dataType to the list if needed + // and return the corresponding response + if ( finalDataType ) { + if ( finalDataType !== dataTypes[ 0 ] ) { + dataTypes.unshift( finalDataType ); + } + return responses[ finalDataType ]; + } +} + +/* Chain conversions given the request and the original response + * Also sets the responseXXX fields on the jqXHR instance + */ +function ajaxConvert( s, response, jqXHR, isSuccess ) { + var conv2, current, conv, tmp, prev, + converters = {}, + + // Work with a copy of dataTypes in case we need to modify it for conversion + dataTypes = s.dataTypes.slice(); + + // Create converters map with lowercased keys + if ( dataTypes[ 1 ] ) { + for ( conv in s.converters ) { + converters[ conv.toLowerCase() ] = s.converters[ conv ]; + } + } + + current = dataTypes.shift(); + + // Convert to each sequential dataType + while ( current ) { + + if ( s.responseFields[ current ] ) { + jqXHR[ s.responseFields[ current ] ] = response; + } + + // Apply the dataFilter if provided + if ( !prev && isSuccess && s.dataFilter ) { + response = s.dataFilter( response, s.dataType ); + } + + prev = current; + current = dataTypes.shift(); + + if ( current ) { + + // There's only work to do if current dataType is non-auto + if ( current === "*" ) { + + current = prev; + + // Convert response if prev dataType is non-auto and differs from current + } else if ( prev !== "*" && prev !== current ) { + + // Seek a direct converter + conv = converters[ prev + " " + current ] || converters[ "* " + current ]; + + // If none found, seek a pair + if ( !conv ) { + for ( conv2 in converters ) { + + // If conv2 outputs current + tmp = conv2.split( " " ); + if ( tmp[ 1 ] === current ) { + + // If prev can be converted to accepted input + conv = converters[ prev + " " + tmp[ 0 ] ] || + converters[ "* " + tmp[ 0 ] ]; + if ( conv ) { + + // Condense equivalence converters + if ( conv === true ) { + conv = converters[ conv2 ]; + + // Otherwise, insert the intermediate dataType + } else if ( converters[ conv2 ] !== true ) { + current = tmp[ 0 ]; + dataTypes.unshift( tmp[ 1 ] ); + } + break; + } + } + } + } + + // Apply converter (if not an equivalence) + if ( conv !== true ) { + + // Unless errors are allowed to bubble, catch and return them + if ( conv && s.throws ) { + response = conv( response ); + } else { + try { + response = conv( response ); + } catch ( e ) { + return { + state: "parsererror", + error: conv ? e : "No conversion from " + prev + " to " + current + }; + } + } + } + } + } + } + + return { state: "success", data: response }; +} + +jQuery.extend( { + + // Counter for holding the number of active queries + active: 0, + + // Last-Modified header cache for next request + lastModified: {}, + etag: {}, + + ajaxSettings: { + url: location.href, + type: "GET", + isLocal: rlocalProtocol.test( location.protocol ), + global: true, + processData: true, + async: true, + contentType: "application/x-www-form-urlencoded; charset=UTF-8", + + /* + timeout: 0, + data: null, + dataType: null, + username: null, + password: null, + cache: null, + throws: false, + traditional: false, + headers: {}, + */ + + accepts: { + "*": allTypes, + text: "text/plain", + html: "text/html", + xml: "application/xml, text/xml", + json: "application/json, text/javascript" + }, + + contents: { + xml: /\bxml\b/, + html: /\bhtml/, + json: /\bjson\b/ + }, + + responseFields: { + xml: "responseXML", + text: "responseText", + json: "responseJSON" + }, + + // Data converters + // Keys separate source (or catchall "*") and destination types with a single space + converters: { + + // Convert anything to text + "* text": String, + + // Text to html (true = no transformation) + "text html": true, + + // Evaluate text as a json expression + "text json": JSON.parse, + + // Parse text as xml + "text xml": jQuery.parseXML + }, + + // For options that shouldn't be deep extended: + // you can add your own custom options here if + // and when you create one that shouldn't be + // deep extended (see ajaxExtend) + flatOptions: { + url: true, + context: true + } + }, + + // Creates a full fledged settings object into target + // with both ajaxSettings and settings fields. + // If target is omitted, writes into ajaxSettings. + ajaxSetup: function( target, settings ) { + return settings ? + + // Building a settings object + ajaxExtend( ajaxExtend( target, jQuery.ajaxSettings ), settings ) : + + // Extending ajaxSettings + ajaxExtend( jQuery.ajaxSettings, target ); + }, + + ajaxPrefilter: addToPrefiltersOrTransports( prefilters ), + ajaxTransport: addToPrefiltersOrTransports( transports ), + + // Main method + ajax: function( url, options ) { + + // If url is an object, simulate pre-1.5 signature + if ( typeof url === "object" ) { + options = url; + url = undefined; + } + + // Force options to be an object + options = options || {}; + + var transport, + + // URL without anti-cache param + cacheURL, + + // Response headers + responseHeadersString, + responseHeaders, + + // timeout handle + timeoutTimer, + + // Url cleanup var + urlAnchor, + + // Request state (becomes false upon send and true upon completion) + completed, + + // To know if global events are to be dispatched + fireGlobals, + + // Loop variable + i, + + // uncached part of the url + uncached, + + // Create the final options object + s = jQuery.ajaxSetup( {}, options ), + + // Callbacks context + callbackContext = s.context || s, + + // Context for global events is callbackContext if it is a DOM node or jQuery collection + globalEventContext = s.context && + ( callbackContext.nodeType || callbackContext.jquery ) ? + jQuery( callbackContext ) : + jQuery.event, + + // Deferreds + deferred = jQuery.Deferred(), + completeDeferred = jQuery.Callbacks( "once memory" ), + + // Status-dependent callbacks + statusCode = s.statusCode || {}, + + // Headers (they are sent all at once) + requestHeaders = {}, + requestHeadersNames = {}, + + // Default abort message + strAbort = "canceled", + + // Fake xhr + jqXHR = { + readyState: 0, + + // Builds headers hashtable if needed + getResponseHeader: function( key ) { + var match; + if ( completed ) { + if ( !responseHeaders ) { + responseHeaders = {}; + while ( ( match = rheaders.exec( responseHeadersString ) ) ) { + responseHeaders[ match[ 1 ].toLowerCase() ] = match[ 2 ]; + } + } + match = responseHeaders[ key.toLowerCase() ]; + } + return match == null ? null : match; + }, + + // Raw string + getAllResponseHeaders: function() { + return completed ? responseHeadersString : null; + }, + + // Caches the header + setRequestHeader: function( name, value ) { + if ( completed == null ) { + name = requestHeadersNames[ name.toLowerCase() ] = + requestHeadersNames[ name.toLowerCase() ] || name; + requestHeaders[ name ] = value; + } + return this; + }, + + // Overrides response content-type header + overrideMimeType: function( type ) { + if ( completed == null ) { + s.mimeType = type; + } + return this; + }, + + // Status-dependent callbacks + statusCode: function( map ) { + var code; + if ( map ) { + if ( completed ) { + + // Execute the appropriate callbacks + jqXHR.always( map[ jqXHR.status ] ); + } else { + + // Lazy-add the new callbacks in a way that preserves old ones + for ( code in map ) { + statusCode[ code ] = [ statusCode[ code ], map[ code ] ]; + } + } + } + return this; + }, + + // Cancel the request + abort: function( statusText ) { + var finalText = statusText || strAbort; + if ( transport ) { + transport.abort( finalText ); + } + done( 0, finalText ); + return this; + } + }; + + // Attach deferreds + deferred.promise( jqXHR ); + + // Add protocol if not provided (prefilters might expect it) + // Handle falsy url in the settings object (#10093: consistency with old signature) + // We also use the url parameter if available + s.url = ( ( url || s.url || location.href ) + "" ) + .replace( rprotocol, location.protocol + "//" ); + + // Alias method option to type as per ticket #12004 + s.type = options.method || options.type || s.method || s.type; + + // Extract dataTypes list + s.dataTypes = ( s.dataType || "*" ).toLowerCase().match( rnothtmlwhite ) || [ "" ]; + + // A cross-domain request is in order when the origin doesn't match the current origin. + if ( s.crossDomain == null ) { + urlAnchor = document.createElement( "a" ); + + // Support: IE <=8 - 11, Edge 12 - 13 + // IE throws exception on accessing the href property if url is malformed, + // e.g. http://example.com:80x/ + try { + urlAnchor.href = s.url; + + // Support: IE <=8 - 11 only + // Anchor's host property isn't correctly set when s.url is relative + urlAnchor.href = urlAnchor.href; + s.crossDomain = originAnchor.protocol + "//" + originAnchor.host !== + urlAnchor.protocol + "//" + urlAnchor.host; + } catch ( e ) { + + // If there is an error parsing the URL, assume it is crossDomain, + // it can be rejected by the transport if it is invalid + s.crossDomain = true; + } + } + + // Convert data if not already a string + if ( s.data && s.processData && typeof s.data !== "string" ) { + s.data = jQuery.param( s.data, s.traditional ); + } + + // Apply prefilters + inspectPrefiltersOrTransports( prefilters, s, options, jqXHR ); + + // If request was aborted inside a prefilter, stop there + if ( completed ) { + return jqXHR; + } + + // We can fire global events as of now if asked to + // Don't fire events if jQuery.event is undefined in an AMD-usage scenario (#15118) + fireGlobals = jQuery.event && s.global; + + // Watch for a new set of requests + if ( fireGlobals && jQuery.active++ === 0 ) { + jQuery.event.trigger( "ajaxStart" ); + } + + // Uppercase the type + s.type = s.type.toUpperCase(); + + // Determine if request has content + s.hasContent = !rnoContent.test( s.type ); + + // Save the URL in case we're toying with the If-Modified-Since + // and/or If-None-Match header later on + // Remove hash to simplify url manipulation + cacheURL = s.url.replace( rhash, "" ); + + // More options handling for requests with no content + if ( !s.hasContent ) { + + // Remember the hash so we can put it back + uncached = s.url.slice( cacheURL.length ); + + // If data is available, append data to url + if ( s.data ) { + cacheURL += ( rquery.test( cacheURL ) ? "&" : "?" ) + s.data; + + // #9682: remove data so that it's not used in an eventual retry + delete s.data; + } + + // Add or update anti-cache param if needed + if ( s.cache === false ) { + cacheURL = cacheURL.replace( rantiCache, "$1" ); + uncached = ( rquery.test( cacheURL ) ? "&" : "?" ) + "_=" + ( nonce++ ) + uncached; + } + + // Put hash and anti-cache on the URL that will be requested (gh-1732) + s.url = cacheURL + uncached; + + // Change '%20' to '+' if this is encoded form body content (gh-2658) + } else if ( s.data && s.processData && + ( s.contentType || "" ).indexOf( "application/x-www-form-urlencoded" ) === 0 ) { + s.data = s.data.replace( r20, "+" ); + } + + // Set the If-Modified-Since and/or If-None-Match header, if in ifModified mode. + if ( s.ifModified ) { + if ( jQuery.lastModified[ cacheURL ] ) { + jqXHR.setRequestHeader( "If-Modified-Since", jQuery.lastModified[ cacheURL ] ); + } + if ( jQuery.etag[ cacheURL ] ) { + jqXHR.setRequestHeader( "If-None-Match", jQuery.etag[ cacheURL ] ); + } + } + + // Set the correct header, if data is being sent + if ( s.data && s.hasContent && s.contentType !== false || options.contentType ) { + jqXHR.setRequestHeader( "Content-Type", s.contentType ); + } + + // Set the Accepts header for the server, depending on the dataType + jqXHR.setRequestHeader( + "Accept", + s.dataTypes[ 0 ] && s.accepts[ s.dataTypes[ 0 ] ] ? + s.accepts[ s.dataTypes[ 0 ] ] + + ( s.dataTypes[ 0 ] !== "*" ? ", " + allTypes + "; q=0.01" : "" ) : + s.accepts[ "*" ] + ); + + // Check for headers option + for ( i in s.headers ) { + jqXHR.setRequestHeader( i, s.headers[ i ] ); + } + + // Allow custom headers/mimetypes and early abort + if ( s.beforeSend && + ( s.beforeSend.call( callbackContext, jqXHR, s ) === false || completed ) ) { + + // Abort if not done already and return + return jqXHR.abort(); + } + + // Aborting is no longer a cancellation + strAbort = "abort"; + + // Install callbacks on deferreds + completeDeferred.add( s.complete ); + jqXHR.done( s.success ); + jqXHR.fail( s.error ); + + // Get transport + transport = inspectPrefiltersOrTransports( transports, s, options, jqXHR ); + + // If no transport, we auto-abort + if ( !transport ) { + done( -1, "No Transport" ); + } else { + jqXHR.readyState = 1; + + // Send global event + if ( fireGlobals ) { + globalEventContext.trigger( "ajaxSend", [ jqXHR, s ] ); + } + + // If request was aborted inside ajaxSend, stop there + if ( completed ) { + return jqXHR; + } + + // Timeout + if ( s.async && s.timeout > 0 ) { + timeoutTimer = window.setTimeout( function() { + jqXHR.abort( "timeout" ); + }, s.timeout ); + } + + try { + completed = false; + transport.send( requestHeaders, done ); + } catch ( e ) { + + // Rethrow post-completion exceptions + if ( completed ) { + throw e; + } + + // Propagate others as results + done( -1, e ); + } + } + + // Callback for when everything is done + function done( status, nativeStatusText, responses, headers ) { + var isSuccess, success, error, response, modified, + statusText = nativeStatusText; + + // Ignore repeat invocations + if ( completed ) { + return; + } + + completed = true; + + // Clear timeout if it exists + if ( timeoutTimer ) { + window.clearTimeout( timeoutTimer ); + } + + // Dereference transport for early garbage collection + // (no matter how long the jqXHR object will be used) + transport = undefined; + + // Cache response headers + responseHeadersString = headers || ""; + + // Set readyState + jqXHR.readyState = status > 0 ? 4 : 0; + + // Determine if successful + isSuccess = status >= 200 && status < 300 || status === 304; + + // Get response data + if ( responses ) { + response = ajaxHandleResponses( s, jqXHR, responses ); + } + + // Convert no matter what (that way responseXXX fields are always set) + response = ajaxConvert( s, response, jqXHR, isSuccess ); + + // If successful, handle type chaining + if ( isSuccess ) { + + // Set the If-Modified-Since and/or If-None-Match header, if in ifModified mode. + if ( s.ifModified ) { + modified = jqXHR.getResponseHeader( "Last-Modified" ); + if ( modified ) { + jQuery.lastModified[ cacheURL ] = modified; + } + modified = jqXHR.getResponseHeader( "etag" ); + if ( modified ) { + jQuery.etag[ cacheURL ] = modified; + } + } + + // if no content + if ( status === 204 || s.type === "HEAD" ) { + statusText = "nocontent"; + + // if not modified + } else if ( status === 304 ) { + statusText = "notmodified"; + + // If we have data, let's convert it + } else { + statusText = response.state; + success = response.data; + error = response.error; + isSuccess = !error; + } + } else { + + // Extract error from statusText and normalize for non-aborts + error = statusText; + if ( status || !statusText ) { + statusText = "error"; + if ( status < 0 ) { + status = 0; + } + } + } + + // Set data for the fake xhr object + jqXHR.status = status; + jqXHR.statusText = ( nativeStatusText || statusText ) + ""; + + // Success/Error + if ( isSuccess ) { + deferred.resolveWith( callbackContext, [ success, statusText, jqXHR ] ); + } else { + deferred.rejectWith( callbackContext, [ jqXHR, statusText, error ] ); + } + + // Status-dependent callbacks + jqXHR.statusCode( statusCode ); + statusCode = undefined; + + if ( fireGlobals ) { + globalEventContext.trigger( isSuccess ? "ajaxSuccess" : "ajaxError", + [ jqXHR, s, isSuccess ? success : error ] ); + } + + // Complete + completeDeferred.fireWith( callbackContext, [ jqXHR, statusText ] ); + + if ( fireGlobals ) { + globalEventContext.trigger( "ajaxComplete", [ jqXHR, s ] ); + + // Handle the global AJAX counter + if ( !( --jQuery.active ) ) { + jQuery.event.trigger( "ajaxStop" ); + } + } + } + + return jqXHR; + }, + + getJSON: function( url, data, callback ) { + return jQuery.get( url, data, callback, "json" ); + }, + + getScript: function( url, callback ) { + return jQuery.get( url, undefined, callback, "script" ); + } +} ); + +jQuery.each( [ "get", "post" ], function( i, method ) { + jQuery[ method ] = function( url, data, callback, type ) { + + // Shift arguments if data argument was omitted + if ( jQuery.isFunction( data ) ) { + type = type || callback; + callback = data; + data = undefined; + } + + // The url can be an options object (which then must have .url) + return jQuery.ajax( jQuery.extend( { + url: url, + type: method, + dataType: type, + data: data, + success: callback + }, jQuery.isPlainObject( url ) && url ) ); + }; +} ); + + +jQuery._evalUrl = function( url ) { + return jQuery.ajax( { + url: url, + + // Make this explicit, since user can override this through ajaxSetup (#11264) + type: "GET", + dataType: "script", + cache: true, + async: false, + global: false, + "throws": true + } ); +}; + + +jQuery.fn.extend( { + wrapAll: function( html ) { + var wrap; + + if ( this[ 0 ] ) { + if ( jQuery.isFunction( html ) ) { + html = html.call( this[ 0 ] ); + } + + // The elements to wrap the target around + wrap = jQuery( html, this[ 0 ].ownerDocument ).eq( 0 ).clone( true ); + + if ( this[ 0 ].parentNode ) { + wrap.insertBefore( this[ 0 ] ); + } + + wrap.map( function() { + var elem = this; + + while ( elem.firstElementChild ) { + elem = elem.firstElementChild; + } + + return elem; + } ).append( this ); + } + + return this; + }, + + wrapInner: function( html ) { + if ( jQuery.isFunction( html ) ) { + return this.each( function( i ) { + jQuery( this ).wrapInner( html.call( this, i ) ); + } ); + } + + return this.each( function() { + var self = jQuery( this ), + contents = self.contents(); + + if ( contents.length ) { + contents.wrapAll( html ); + + } else { + self.append( html ); + } + } ); + }, + + wrap: function( html ) { + var isFunction = jQuery.isFunction( html ); + + return this.each( function( i ) { + jQuery( this ).wrapAll( isFunction ? html.call( this, i ) : html ); + } ); + }, + + unwrap: function( selector ) { + this.parent( selector ).not( "body" ).each( function() { + jQuery( this ).replaceWith( this.childNodes ); + } ); + return this; + } +} ); + + +jQuery.expr.pseudos.hidden = function( elem ) { + return !jQuery.expr.pseudos.visible( elem ); +}; +jQuery.expr.pseudos.visible = function( elem ) { + return !!( elem.offsetWidth || elem.offsetHeight || elem.getClientRects().length ); +}; + + + + +jQuery.ajaxSettings.xhr = function() { + try { + return new window.XMLHttpRequest(); + } catch ( e ) {} +}; + +var xhrSuccessStatus = { + + // File protocol always yields status code 0, assume 200 + 0: 200, + + // Support: IE <=9 only + // #1450: sometimes IE returns 1223 when it should be 204 + 1223: 204 + }, + xhrSupported = jQuery.ajaxSettings.xhr(); + +support.cors = !!xhrSupported && ( "withCredentials" in xhrSupported ); +support.ajax = xhrSupported = !!xhrSupported; + +jQuery.ajaxTransport( function( options ) { + var callback, errorCallback; + + // Cross domain only allowed if supported through XMLHttpRequest + if ( support.cors || xhrSupported && !options.crossDomain ) { + return { + send: function( headers, complete ) { + var i, + xhr = options.xhr(); + + xhr.open( + options.type, + options.url, + options.async, + options.username, + options.password + ); + + // Apply custom fields if provided + if ( options.xhrFields ) { + for ( i in options.xhrFields ) { + xhr[ i ] = options.xhrFields[ i ]; + } + } + + // Override mime type if needed + if ( options.mimeType && xhr.overrideMimeType ) { + xhr.overrideMimeType( options.mimeType ); + } + + // X-Requested-With header + // For cross-domain requests, seeing as conditions for a preflight are + // akin to a jigsaw puzzle, we simply never set it to be sure. + // (it can always be set on a per-request basis or even using ajaxSetup) + // For same-domain requests, won't change header if already provided. + if ( !options.crossDomain && !headers[ "X-Requested-With" ] ) { + headers[ "X-Requested-With" ] = "XMLHttpRequest"; + } + + // Set headers + for ( i in headers ) { + xhr.setRequestHeader( i, headers[ i ] ); + } + + // Callback + callback = function( type ) { + return function() { + if ( callback ) { + callback = errorCallback = xhr.onload = + xhr.onerror = xhr.onabort = xhr.onreadystatechange = null; + + if ( type === "abort" ) { + xhr.abort(); + } else if ( type === "error" ) { + + // Support: IE <=9 only + // On a manual native abort, IE9 throws + // errors on any property access that is not readyState + if ( typeof xhr.status !== "number" ) { + complete( 0, "error" ); + } else { + complete( + + // File: protocol always yields status 0; see #8605, #14207 + xhr.status, + xhr.statusText + ); + } + } else { + complete( + xhrSuccessStatus[ xhr.status ] || xhr.status, + xhr.statusText, + + // Support: IE <=9 only + // IE9 has no XHR2 but throws on binary (trac-11426) + // For XHR2 non-text, let the caller handle it (gh-2498) + ( xhr.responseType || "text" ) !== "text" || + typeof xhr.responseText !== "string" ? + { binary: xhr.response } : + { text: xhr.responseText }, + xhr.getAllResponseHeaders() + ); + } + } + }; + }; + + // Listen to events + xhr.onload = callback(); + errorCallback = xhr.onerror = callback( "error" ); + + // Support: IE 9 only + // Use onreadystatechange to replace onabort + // to handle uncaught aborts + if ( xhr.onabort !== undefined ) { + xhr.onabort = errorCallback; + } else { + xhr.onreadystatechange = function() { + + // Check readyState before timeout as it changes + if ( xhr.readyState === 4 ) { + + // Allow onerror to be called first, + // but that will not handle a native abort + // Also, save errorCallback to a variable + // as xhr.onerror cannot be accessed + window.setTimeout( function() { + if ( callback ) { + errorCallback(); + } + } ); + } + }; + } + + // Create the abort callback + callback = callback( "abort" ); + + try { + + // Do send the request (this may raise an exception) + xhr.send( options.hasContent && options.data || null ); + } catch ( e ) { + + // #14683: Only rethrow if this hasn't been notified as an error yet + if ( callback ) { + throw e; + } + } + }, + + abort: function() { + if ( callback ) { + callback(); + } + } + }; + } +} ); + + + + +// Prevent auto-execution of scripts when no explicit dataType was provided (See gh-2432) +jQuery.ajaxPrefilter( function( s ) { + if ( s.crossDomain ) { + s.contents.script = false; + } +} ); + +// Install script dataType +jQuery.ajaxSetup( { + accepts: { + script: "text/javascript, application/javascript, " + + "application/ecmascript, application/x-ecmascript" + }, + contents: { + script: /\b(?:java|ecma)script\b/ + }, + converters: { + "text script": function( text ) { + jQuery.globalEval( text ); + return text; + } + } +} ); + +// Handle cache's special case and crossDomain +jQuery.ajaxPrefilter( "script", function( s ) { + if ( s.cache === undefined ) { + s.cache = false; + } + if ( s.crossDomain ) { + s.type = "GET"; + } +} ); + +// Bind script tag hack transport +jQuery.ajaxTransport( "script", function( s ) { + + // This transport only deals with cross domain requests + if ( s.crossDomain ) { + var script, callback; + return { + send: function( _, complete ) { + script = jQuery( " + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.change

+

The change is fix incorrect input language correction.

+
+

Modules

+
+
+pythainlp.change.texttothai(data)[source]
+
+++ + + + + + +
Parameters:data (str) – Incorrect input language correction (Needs thai but input english)
Returns:thai text
+
+ +
+
+pythainlp.change.texttoeng(data)[source]
+
+++ + + + + + +
Parameters:data (str) – Incorrect input language correction (Needs english but input thai)
Returns:english text
+
+ +
+
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/collation.html b/docs/_build/html/api/collation.html new file mode 100644 index 000000000..14c68ea4e --- /dev/null +++ b/docs/_build/html/api/collation.html @@ -0,0 +1,261 @@ + + + + + + + + + + + pythainlp.collation — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.collation

+

The pythainlp.collation contains a function that sorts Thai text alphabetically

+
+
+pythainlp.collation.collation(data)[source]
+
+++ + + + + + +
Parameters:data (list) – a list of thai text
Returns:a list of thai text, sorted alphabetically
+
+
Example::
+
>>> from pythainlp.collation import *
+>>> collation(['ไก่', 'เป็ด', 'หมู', 'วัว'])
+['ไก่', 'เป็ด', 'วัว', 'หมู']
+
+
+
+
+
+ +
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/date.html b/docs/_build/html/api/date.html new file mode 100644 index 000000000..ede3eb8ec --- /dev/null +++ b/docs/_build/html/api/date.html @@ -0,0 +1,250 @@ + + + + + + + + + + + pythainlp.date — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.date

+

The pythainlp.date helps output dates, as spelled out in Thai.

+
+
+pythainlp.date.now()[source]
+
+++ + + + +
Returns:the current date with Thai month and Thai year. The month is spelled out in text, and the year is converted from AD to Thai years. (ie: 30 ตุลาคม 2560 20:45:30)
+
+ +
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/ner.html b/docs/_build/html/api/ner.html new file mode 100644 index 000000000..2cf080073 --- /dev/null +++ b/docs/_build/html/api/ner.html @@ -0,0 +1,281 @@ + + + + + + + + + + + pythainlp.ner — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.ner

+

The pythainlp.ner is named entity recognition for thai.

+
+

Modules

+
+
+class pythainlp.ner.thainer[source]
+
+
+get_ner(text, postag=True)[source]
+

Get NER from Thai NER.

+ +++ + + + + + +
Parameters:
    +
  • text (string) – thai text
  • +
  • postag (boolean) – get postag (True) or get not postag (False)
  • +
+
Returns:

list NER.

+
+
+
Example::
+
>>> from pythainlp.ner import thainer
+>>> ner=thainer()
+>>> ner.get_ner("วันที่ 15 ก.ย. 61 ทดสอบระบบเวลา 14:49 น.")
+[('วันที่', 'JSBR', 'O'), (' ', 'NCMN', 'O'), ('15', 'NCNM', 'B-DATE'), (' ', 'NCMN', 'I-DATE'), ('ก.ย.', 'CMTR', 'I-DATE'), (' ', 'NCMN', 'I-DATE'), ('61', 'NCNM', 'I-DATE'), (' ', 'NCMN', 'O'), ('ทดสอบ', 'VACT', 'O'), ('ระบบ', 'NCMN', 'O'), ('เวลา', 'NCMN', 'O'), (' ', 'NCMN', 'O'), ('14', 'NCNM', 'B-TIME'), (':', 'PUNC', 'I-TIME'), ('49', 'NCNM', 'I-TIME'), (' ', 'NCMN', 'I-TIME'), ('น.', 'CMTR', 'I-TIME')]
+>>> ner.get_ner("วันที่ 15 ก.ย. 61 ทดสอบระบบเวลา 14:49 น.",postag=False)
+[('วันที่', 'O'), (' ', 'O'), ('15', 'B-DATE'), (' ', 'I-DATE'), ('ก.ย.', 'I-DATE'), (' ', 'I-DATE'), ('61', 'I-DATE'), (' ', 'O'), ('ทดสอบ', 'O'), ('ระบบ', 'O'), ('เวลา', 'O'), (' ', 'O'), ('14', 'B-TIME'), (':', 'I-TIME'), ('49', 'I-TIME'), (' ', 'I-TIME'), ('น.', 'I-TIME')]
+
+
+
+
+
+ +
+ +
+
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/number.html b/docs/_build/html/api/number.html new file mode 100644 index 000000000..d6cad6dc3 --- /dev/null +++ b/docs/_build/html/api/number.html @@ -0,0 +1,317 @@ + + + + + + + + + + + pythainlp.number — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.number

+

The pythainlp.number contains functions for processing thai numbers and thai words that refers to numbers.

+
+
+pythainlp.number.thai_num_to_num(text)[source]
+
+++ + + + + + +
Parameters:text (str) – Thai number characters such as ‘๑’, ‘๒’, ‘๓’
Returns:universal numbers such as ‘1’, ‘2’, ‘3’
+
+ +
+
+pythainlp.number.thai_num_to_text(text)[source]
+
+++ + + + + + +
Parameters:text (str) – Thai number characters such as ‘๑’, ‘๒’, ‘๓’
Returns:Thai numbers, spelled out in Thai
+
+ +
+
+pythainlp.number.num_to_thai_num(text)[source]
+
+++ + + + + + +
Parameters:text – universal numbers such as ‘1’, ‘2’, ‘3’
Returns:Thai number characters such as ‘๑’, ‘๒’, ‘๓’
+
+ +
+
+pythainlp.number.num_to_text(text)[source]
+
+++ + + + + + +
Parameters:text – universal numbers such as ‘1’, ‘2’, ‘3’
Returns:Thai numbers, spelled out in Thai
+
+ +
+
+pythainlp.number.text_to_num(text)[source]
+
+++ + + + + + +
Parameters:text – Thai numbers, spelled out in Thai
Returns:universal numbers such as ‘1’, ‘2’, ‘3’
+
+ +
+
+pythainlp.number.numtowords(amount_number)[source]
+
+ +
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/romanization.html b/docs/_build/html/api/romanization.html new file mode 100644 index 000000000..5977d4bcf --- /dev/null +++ b/docs/_build/html/api/romanization.html @@ -0,0 +1,277 @@ + + + + + + + + + + + pythainlp.romanization — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.romanization

+

The romanization.romanization turns thai text into a romanized one (put simply, spelled with English).

+
+
+pythainlp.romanization.romanization(data, engine='royin')[source]
+
+++ + + + + + +
Parameters:
    +
  • data (str) – Thai text to be romanized
  • +
  • engine (str) – choose between ‘royin’ , ‘pyicu’ and ‘thai2rom’. ‘royin’ will romanize according to the standard of Thai Royal Institute. ‘pyicu’ will romanize according to the Internaitonal Phonetic Alphabet. ‘thai2rom’ is deep learning thai romanization.
  • +
+
Returns:

English (more or less) text that spells out how the Thai text should read.

+
+
+ +
+
+class pythainlp.romanization.thai2rom.thai2rom[source]
+
+
+romanization(text)[source]
+
+++ + + + + + +
Parameters:text (str) – Thai text to be romanized
Returns:English (more or less) text that spells out how the Thai text should read.
+
+ +
+ +
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/sentiment.html b/docs/_build/html/api/sentiment.html new file mode 100644 index 000000000..4854927a2 --- /dev/null +++ b/docs/_build/html/api/sentiment.html @@ -0,0 +1,274 @@ + + + + + + + + + + + pythainlp.sentiment — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.sentiment

+

The romanization.sentiment is sentiment analysis.

+
+
+pythainlp.sentiment.sentiment(text, engine='old')[source]
+
+++ + + + + + +
Parameters:
    +
  • text (str) – thai text
  • +
  • engine (str) – sentiment analysis engine (old or ulmfit)
  • +
+
Returns:

pos or neg

+
+
+
Example::
+
>>> from pythainlp.sentiment import sentiment
+>>> text="วันนี้อากาศดีจัง"
+>>> sentiment(text)
+'pos'
+>>> sentiment(text,'ulmfit')
+'pos'
+>>> text="วันนี้อารมณ์เสียมาก"
+>>> sentiment(text)
+'neg'
+>>> sentiment(text,'ulmfit')
+'neg'
+
+
+
+
+
+ +
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/soundex.html b/docs/_build/html/api/soundex.html new file mode 100644 index 000000000..d81144888 --- /dev/null +++ b/docs/_build/html/api/soundex.html @@ -0,0 +1,269 @@ + + + + + + + + + + + pythainlp.soundex — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.soundex

+

The pythainlp.soundex is soundex for thai.

+
+
+pythainlp.soundex.LK82(s)[source]
+

LK82 - It’s a thai soundex rule.

+ +++ + + + + + +
Parameters:s (str) – thai word
Returns:LK82 soundex
+
+ +
+
+pythainlp.soundex.Udom83(s)[source]
+

Udom83 - It’s a thai soundex rule.

+ +++ + + + + + +
Parameters:s (str) – thai word
Returns:LK82 soundex
+
+ +
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/spell.html b/docs/_build/html/api/spell.html new file mode 100644 index 000000000..45a0987ab --- /dev/null +++ b/docs/_build/html/api/spell.html @@ -0,0 +1,261 @@ + + + + + + + + + + + pythainlp.spell — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.spell

+

The pythainlp.spell finds the closest correctly spelled word to the given text.

+
+
+pythainlp.spell.spell(word, engine='pn')[source]
+
+++ + + + + + +
Parameters:
    +
  • word (str) – the word to check spelling
  • +
  • engine (str) –
      +
    • pn - Peter Norvig’s algorithm
    • +
    • hunspell - uses hunspell’s algorithm, which should already exist in linux
    • +
    +
  • +
+
Returns:

list word

+
+
+ +
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/summarize.html b/docs/_build/html/api/summarize.html new file mode 100644 index 000000000..1af199829 --- /dev/null +++ b/docs/_build/html/api/summarize.html @@ -0,0 +1,253 @@ + + + + + + + + + + + pythainlp.summarize — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.summarize

+

The summarize is thai text summarize.

+
+

Modules

+
+
+pythainlp.summarize.summarize_text(text, n, engine='frequency', tokenize='newmm')[source]
+

Thai text summarize. +:param str text: thai text +:param int n: sent number +:param str engine: Thai text summarize engine. +:param str tokenize: thai word tokenize.

+
+ +
+
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/tag.html b/docs/_build/html/api/tag.html new file mode 100644 index 000000000..ccafd27a1 --- /dev/null +++ b/docs/_build/html/api/tag.html @@ -0,0 +1,268 @@ + + + + + + + + + + + pythainlp.tag — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.tag

+

The pythainlp.tag contains functions that are used to tag different parts of a text.

+
+
+pythainlp.tag.pos_tag(list_text, engine='unigram', corpus='orchid')[source]
+

Part of Speech tagging function.

+ +++ + + + + + +
Parameters:
    +
  • list_text (list) – takes in a list of tokenized words (put differently, a list of string)
  • +
  • engine (str) –
      +
    • unigram - unigram tagger
    • +
    • perceptron - perceptron tagger
    • +
    • artagger - RDR POS tagger
    • +
    +
  • +
  • corpus (str) –
      +
    • orchid - annotated Thai academic articles
    • +
    • pud - Parallel Universal Dependencies (PUD) treebanks
    • +
    +
  • +
+
Returns:

returns a list of labels regarding which part of speech it is

+
+
+ +
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/tokenizer.html b/docs/_build/html/api/tokenizer.html new file mode 100644 index 000000000..2463aabe2 --- /dev/null +++ b/docs/_build/html/api/tokenizer.html @@ -0,0 +1,399 @@ + + + + + + + + + + + pythainlp.tokenize — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.tokenize

+

The pythainlp.tokenize contains multiple functions for tokenizing a chunk of Thai text into desirable units.

+
+
+pythainlp.tokenize.word_tokenize(text, engine='newmm', whitespaces=True)[source]
+
+++ + + + + + + + + +
Parameters:
    +
  • text (str) – the text to be tokenized
  • +
  • engine (str) – the engine to tokenize text
  • +
  • whitespaces (bool) – True to output no whitespace, a common mark of sentence or end of phrase in Thai.
  • +
+
Parameters for engine:
 
    +
  • newmm - Maximum Matching algorithm + TCC
  • +
  • icu - IBM ICU
  • +
  • longest-matching - Longest matching
  • +
  • mm - Maximum Matching algorithm
  • +
  • pylexto - LexTo
  • +
  • deepcut - Deep Neural Network
  • +
  • wordcutpy - wordcutpy (https://github.com/veer66/wordcutpy)
  • +
+
Returns:

A list of words, tokenized from a text

+
+

Example:

+
from pythainlp.tokenize import word_tokenize
+text='ผมรักคุณนะครับโอเคบ่พวกเราเป็นคนไทยรักภาษาไทยภาษาบ้านเกิด'
+a=word_tokenize(text,engine='icu') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอ', 'เค', 'บ่', 'พวก', 'เรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้าน', 'เกิด']
+b=word_tokenize(text,engine='dict') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด']
+c=word_tokenize(text,engine='mm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด']
+d=word_tokenize(text,engine='pylexto') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด']
+e=word_tokenize(text,engine='newmm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด']
+g=word_tokenize(text,engine='wordcutpy') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้านเกิด']
+
+
+
+ +
+
+pythainlp.tokenize.dict_word_tokenize(text, custom_dict_trie, engine='newmm')[source]
+
+

dict_word_tokenize() tokenizes word based on the dictionary you provide. The format has to be in trie data structure.

+ +++ + + + + + + + + + +
param str text:the text to be tokenized
param dict custom_dict_trie:
 คือ trie ที่สร้างจาก create_custom_dict_trie
param str engine:
 choose between different options of engine to token (newmm, wordcutpy, mm, longest-matching)
+
+ +++ + + + +
Returns:A list of words, tokenized from a text.
+
+
Example::
+
>>> from pythainlp.tokenize import dict_word_tokenize,create_custom_dict_trie
+>>> listword=['แมว',"ดี"]
+>>> data_dict=create_custom_dict_trie(listword)
+>>> dict_word_tokenize("แมวดีดีแมว",data_dict)
+['แมว', 'ดี', 'ดี', 'แมว']
+
+
+
+
+
+ +
+
+pythainlp.tokenize.subword_tokenize(text, engine='tcc')[source]
+
+++ + + + + + +
Parameters:
    +
  • text (str) – text to be tokenized
  • +
  • engine (str) – choosing ‘tcc’ uses the Thai Character Cluster rule to segment words into the smallest unique units.
  • +
+
Returns:

a list of tokenized strings.

+
+
+ +
+
+pythainlp.tokenize.sent_tokenize(text, engine='whitespace+newline')[source]
+

This function does not yet automatically recognize when a sentence actually ends. Rather it helps split text where white space and a new line is found.

+ +++ + + + + + +
Parameters:
    +
  • text (str) – the text to be tokenized
  • +
  • engine (str) – choose between ‘whitespace’ or ‘whitespace+newline’
  • +
+
Returns:

a list of text, split by whitespace or new line.

+
+
+ +
+
+pythainlp.tokenize.isthai(text, check_all=False)[source]
+
+++ + + + + + +
Parameters:
    +
  • text (str) – input string or list of strings
  • +
  • check_all (bool) – checks all character or not
  • +
+
Returns:

A dictionary with the first value as proportional of text that is Thai, and the second value being a tuple of all characters, along with true or false.

+
+
+ +
+
+pythainlp.tokenize.create_custom_dict_trie(custom_dict_source)[source]
+

The function is used to create a custom dict trie which will be used for word_tokenize() function. For more information on the trie data structure, see: https://marisa-trie.readthedocs.io/en/latest/index.html

+ +++ + + + + + +
Parameters:custom_dict_source (string/list) – a list of vocaburaries or a path to source file
Returns:A trie created from custom dict input
+
+ +
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/ulmfit.html b/docs/_build/html/api/ulmfit.html new file mode 100644 index 000000000..ba59e631e --- /dev/null +++ b/docs/_build/html/api/ulmfit.html @@ -0,0 +1,486 @@ + + + + + + + + + + + pythainlp.ulmfit — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.ulmfit

+

The ulmfit.utils is utils for ULMFit model.

+
+

Modules

+
+
+pythainlp.ulmfit.utils.get_texts(df)[source]
+
+++ + + + + + + + +
Meth:get_texts get tuple of tokenized texts and labels
Parameters:df (pandas.DataFrame) – pandas.DataFrame with label as first column and text as second column
Returns:
    +
  • tok - lists of tokenized texts with beginning-of-sentence tag xbos as first element of each list
  • +
  • labels - list of labels
  • +
+
+
+ +
+
+pythainlp.ulmfit.utils.get_all(df)[source]
+
+++ + + + + + + + +
Meth:get_all iterate get_texts for all the entire pandas.DataFrame
Parameters:df (pandas.DataFrame) – pandas.DataFrame with label as first column and text as second column
Returns:
    +
  • tok - lists of tokenized texts with beginning-of-sentence tag xbos as first element of each list
  • +
  • labels - list of labels
  • +
+
+
+ +
+
+pythainlp.ulmfit.utils.numericalizer(df, itos=None, max_vocab=60000, min_freq=2, pad_tok='_pad_', unk_tok='_unk_')[source]
+
+++ + + + + + + + +
Meth:

numericalize numericalize tokenized texts for: +* tokens with word frequency more than min_freq +* at maximum vocab size of max_vocab +* add unknown token _unk_ and padding token _pad_ in first and second position +* use integer-to-string list itos if avaiable e.g. [‘_unk_’, ‘_pad_’,’first_word’,’second_word’,…]

+
Parameters:
    +
  • df (pandas.DataFrame) – pandas.DataFrame with label as first column and text as second column
  • +
  • itos (list) – integer-to-string list
  • +
  • max_vocab (int) – maximum number of vocabulary (default 60000)
  • +
  • min_freq (int) – minimum word frequency to be included (default 2)
  • +
  • pad_tok (str) – padding token
  • +
  • unk_token (str) – unknown token
  • +
+
Returns:

    +
  • lm - numpy.array of numericalized texts
  • +
  • tok - lists of tokenized texts with beginning-of-sentence tag xbos as first element of each list
  • +
  • labels - list of labels
  • +
  • itos - integer-to-string list e.g. [‘_unk_’, ‘_pad_’,’first_word’,’second_word’,…]
  • +
  • stoi - string-to-integer dict e.g. {‘_unk_’:0, ‘_pad_’:1,’first_word’:2,’second_word’:3,…}
  • +
  • freq - collections.Counter for word frequency
  • +
+

+
+
+ +
+
+pythainlp.ulmfit.utils.merge_wgts(em_sz, wgts, itos_pre, itos_cls)[source]
+
+++ + + + + + +
Parameters:
    +
  • df (pandas.DataFrame) – pandas.DataFrame with label as first column and text as second column
  • +
  • em_sz (int) – size of embedding vectors (pretrained model is at 300)
  • +
  • wgts – saved pyTorch weights of pretrained model
  • +
  • itos_pre (list) – integer-to-string list of pretrained model
  • +
  • itos_cls (list) – integer-to-string list of current dataset
  • +
+
Returns:

merged weights of the model for current dataset

+
+
+ +
+
+pythainlp.ulmfit.utils.document_vector(ss, m, stoi, tok_engine='newmm')[source]
+
+++ + + + + + + + +
Meth:

document_vector get document vector using pretrained ULMFit model

+
Parameters:
    +
  • ss (str) – sentence to extract embeddings
  • +
  • m – pyTorch model
  • +
  • stoi (dict) – string-to-integer dict e.g. {‘_unk_’:0, ‘_pad_’:1,’first_word’:2,’second_word’:3,…}
  • +
  • tok_engine (str) – tokenization engine (recommend using newmm if you are using pretrained ULMFit model)
  • +
+
Returns:

numpy.array of document vector sized 300

+
+
+ +
+
+pythainlp.ulmfit.utils.about()[source]
+
+ +
+
+class pythainlp.ulmfit.utils.ThaiTokenizer(engine='newmm')[source]
+
+
+static proc_all(ss)[source]
+
+++ + + + + + + + +
Meth:proc_all runs proc_text for multiple sentences
Parameters:text (str) – text to process
Returns:processed and tokenized text
+
+ +
+
+static proc_all_mp(ss)[source]
+
+++ + + + + + + + +
Meth:proc_all runs proc_text for multiple sentences using multiple cpus
Parameters:text (str) – text to process
Returns:processed and tokenized text
+
+ +
+
+proc_text(text)[source]
+
+++ + + + + + + + +
Meth:proc_text procss and tokenize text removing repetitions, special characters, double spaces
Parameters:text (str) – text to process
Returns:processed and tokenized text
+
+ +
+
+static replace_rep(text)[source]
+

replace_rep() replace 3 or above repetitive characters with tkrep +:param str text: text to process +:return: processed text where repetitions are replaced by tkrep followed by number of repetitions +Example:

+
>>> from pythainlp.ulmfit.utils import ThaiTokenizer
+>>> tt = ThaiTokenizer()
+>>> tt.replace_rep('คือดียยยยยย')
+คือดีtkrep6ย
+
+
+
+ +
+
+sub_br(text)[source]
+
+
sub_br() replace <br> tags with `
+
+
`
+
+++ + + + + + +
param str text:text to process
return:procssed text
+
+
+
+ +
+
+tokenize(text)[source]
+
+++ + + + + + + + +
Meth:tokenize text with selected engine
Parameters:text (str) – text to tokenize
Returns:tokenized text
+
+ +
+ +
+
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/api/word_vector.html b/docs/_build/html/api/word_vector.html new file mode 100644 index 000000000..fd2269028 --- /dev/null +++ b/docs/_build/html/api/word_vector.html @@ -0,0 +1,300 @@ + + + + + + + + + + + pythainlp.word_vector — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

pythainlp.word_vector

+

The word_vector.thai2vec contains functions that makes use of a pre-trained vector public data.

+
+

Dependencies

+

Installation of numpy and gensim is required.

+
+
+

Modules

+
+
+pythainlp.word_vector.thai2vec.get_model()[source]
+
+++ + + + +
Returns:Downloads the gensim model.
+
+ +
+
+pythainlp.word_vector.thai2vec.most_similar_cosmul(positive, negative)[source]
+

การใช้งาน +input list

+
+ +
+
+pythainlp.word_vector.thai2vec.doesnt_match(listdata)[source]
+
+ +
+
+pythainlp.word_vector.thai2vec.similarity(word1, word2)[source]
+
+++ + + + + + +
Parameters:
    +
  • word1 (str) – first word
  • +
  • word2 (str) – second word
  • +
+
Returns:

the cosine similarity between the two word vectors

+
+
+ +
+
+pythainlp.word_vector.thai2vec.sentence_vectorizer(ss, dim=300, use_mean=False)[source]
+
+ +
+
+pythainlp.word_vector.thai2vec.about()[source]
+
+ +
+
+ + +
+ +
+
+ + + + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/genindex.html b/docs/_build/html/genindex.html new file mode 100644 index 000000000..dfe083b93 --- /dev/null +++ b/docs/_build/html/genindex.html @@ -0,0 +1,439 @@ + + + + + + + + + + + + Index — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ +
    + +
  • Docs »
  • + +
  • Index
  • + + +
  • + + + +
  • + +
+ + +
+
+
+
+ + +

Index

+ +
+ A + | C + | D + | G + | I + | L + | M + | N + | P + | R + | S + | T + | U + | W + +
+

A

+ + +
+ +

C

+ + + +
+ +

D

+ + + +
+ +

G

+ + + +
+ +

I

+ + +
+ +

L

+ + +
+ +

M

+ + + +
+ +

N

+ + + +
+ +

P

+ + + +
+ +

R

+ + + +
+ +

S

+ + + +
+ +

T

+ + + +
+ +

U

+ + +
+ +

W

+ + +
+ + + +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/index.html b/docs/_build/html/index.html new file mode 100644 index 000000000..adafe8ff1 --- /dev/null +++ b/docs/_build/html/index.html @@ -0,0 +1,270 @@ + + + + + + + + + + + PyThaiNLP documentation — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

PyThaiNLP documentation

+

PyThaiNLP is a Python library for natural language processing (NLP) of Thai language.

+ + +
+
+

Indices and tables

+ +

Apache Software License 2.0

+

Maintained by the PyThaiNLP team.

+
+ + +
+ +
+
+ + + + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/notes/getting_started.html b/docs/_build/html/notes/getting_started.html new file mode 100644 index 000000000..e64a813af --- /dev/null +++ b/docs/_build/html/notes/getting_started.html @@ -0,0 +1,255 @@ + + + + + + + + + + + Getting Started — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

Getting Started

+

PyThaiNLP is a Python library for natural language processing (NLP) of Thai language. With this package, you can NLP tasks such as classify texts and tokenize words.

+

Tokenization Example:

+
from pythainlp.tokenize import word_tokenize
+text='ผมรักคุณนะครับโอเคบ่พวกเราเป็นคนไทยรักภาษาไทยภาษาบ้านเกิด'
+a=word_tokenize(text,engine='icu') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอ', 'เค', 'บ่', 'พวก', 'เรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้าน', 'เกิด']
+b=word_tokenize(text,engine='dict') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด']
+c=word_tokenize(text,engine='mm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด']
+d=word_tokenize(text,engine='pylexto') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด']
+e=word_tokenize(text,engine='newmm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด']
+g=word_tokenize(text,engine='wordcutpy') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้านเกิ
+
+
+

Thai has historically faced a lot of NLP challenges. A quick list of them include as follows:

+
    +
  1. Start-end of sentence marking - This is arguably the biggest problem for the field of Thai NLP. The lack of end of sentence marking (EOS) makes it hard for researchers to create training sets, the basis of most research in this field. The root of the problem is two-pronged. In terms of writing system, Thai uses space to indicate both commas and periods. No letter indicates an end of a sentence. In terms of language use, Thais have a habit of starting their sentences with connector terms such as ‘because’, ‘but’, ‘following’, etc, making it often hard even for natives to decide where the end of sentence should be when translating.
  2. +
  3. Word segmentation - Thai does not use space and word segmentation is not easy. It boils down to understanding the context and ruling out words that do not make sense. This is a similar issue that other Asian languages such as Japanese and Chinese face in different degrees. For languages with space, a similar but less extreme problem would be multi-word expressions, like the French word for potato — ‘pomme de terre’. In Thai, the best known example is “ตา-กลม” and “ตาก-ลม”. As of recent, new techniques that capture words, subwords, and letters in vectors seem poised to overcome to issue.
  4. +
+

Example for classification

+

https://colab.research.google.com/drive/1cnJ6O3b1jwaHwvsMWW3oQw7f8X2Ka7Sp

+
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/notes/installation.html b/docs/_build/html/notes/installation.html new file mode 100644 index 000000000..da296b2e2 --- /dev/null +++ b/docs/_build/html/notes/installation.html @@ -0,0 +1,245 @@ + + + + + + + + + + + Installation — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ + + + +
+
+
+
+ +
+

Installation

+

For stable version, try:

+
pip install pythainlp
+
+
+

For developer version, try:

+
pip install https://github.com/PyThaiNLP/pythainlp/archive/dev.zip
+
+
+

Note for Windows: marisa-trie wheels can be obtained from https://www.lfd.uci.edu/~gohlke/pythonlibs/#marisa-trie , then install it with pip, for example: pip install marisa_trie‑0.7.5‑cp36‑cp36m‑win32.whl

+
+ + +
+ +
+ + +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/objects.inv b/docs/_build/html/objects.inv new file mode 100644 index 000000000..53ffcdc36 --- /dev/null +++ b/docs/_build/html/objects.inv @@ -0,0 +1,12 @@ +# Sphinx inventory version 2 +# Project: PyThaiNLP +# Version: +# The remainder of this file is compressed using zlib. +xڭWMo0 WخR ú@Ȭ#L$MȎ-YmǗX!iK۽QD[U T^Ӡ; +Ŵ_gAm4DUJ*@PLPk{Hp}pGJp;]vr=B&=2NýB +_tH;0%i1\ + + + + + + + + Search — PyThaiNLP 1.7 documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + +
+ + + + + +
+ +
+ + + + + + + + + + + + + + + + + +
+ +
    + +
  • Docs »
  • + +
  • Search
  • + + +
  • + + + +
  • + +
+ + +
+
+
+
+ + + + +
+ +
+ +
+ +
+
+ + +
+ +
+

+ © Copyright 2017-2018, PyThaiNLP (Apache Software License 2.0). + +

+
+ Built with Sphinx using a theme provided by Read the Docs. + +
+ +
+
+ +
+ +
+ + + + + + + + + + + + + + + + + + + + + + + + + \ No newline at end of file diff --git a/docs/_build/html/searchindex.js b/docs/_build/html/searchindex.js new file mode 100644 index 000000000..d272b6716 --- /dev/null +++ b/docs/_build/html/searchindex.js @@ -0,0 +1 @@ +Search.setIndex({docnames:["api/change","api/collation","api/date","api/ner","api/number","api/romanization","api/sentiment","api/soundex","api/spell","api/summarize","api/tag","api/tokenizer","api/ulmfit","api/word_vector","index","notes/getting_started","notes/installation"],envversion:55,filenames:["api/change.rst","api/collation.rst","api/date.rst","api/ner.rst","api/number.rst","api/romanization.rst","api/sentiment.rst","api/soundex.rst","api/spell.rst","api/summarize.rst","api/tag.rst","api/tokenizer.rst","api/ulmfit.rst","api/word_vector.rst","index.rst","notes/getting_started.rst","notes/installation.rst"],objects:{"pythainlp.change":{texttoeng:[0,0,1,""],texttothai:[0,0,1,""]},"pythainlp.collation":{collation:[1,0,1,""]},"pythainlp.date":{now:[2,0,1,""]},"pythainlp.ner":{thainer:[3,1,1,""]},"pythainlp.ner.thainer":{get_ner:[3,2,1,""]},"pythainlp.number":{num_to_text:[4,0,1,""],num_to_thai_num:[4,0,1,""],numtowords:[4,0,1,""],text_to_num:[4,0,1,""],thai_num_to_num:[4,0,1,""],thai_num_to_text:[4,0,1,""]},"pythainlp.romanization":{romanization:[5,0,1,""]},"pythainlp.romanization.thai2rom":{thai2rom:[5,1,1,""]},"pythainlp.romanization.thai2rom.thai2rom":{romanization:[5,2,1,""]},"pythainlp.sentiment":{sentiment:[6,0,1,""]},"pythainlp.soundex":{LK82:[7,0,1,""],Udom83:[7,0,1,""]},"pythainlp.spell":{spell:[8,0,1,""]},"pythainlp.summarize":{summarize_text:[9,0,1,""]},"pythainlp.tag":{pos_tag:[10,0,1,""]},"pythainlp.tokenize":{create_custom_dict_trie:[11,0,1,""],dict_word_tokenize:[11,0,1,""],isthai:[11,0,1,""],sent_tokenize:[11,0,1,""],subword_tokenize:[11,0,1,""],word_tokenize:[11,0,1,""]},"pythainlp.ulmfit.utils":{ThaiTokenizer:[12,1,1,""],about:[12,0,1,""],document_vector:[12,0,1,""],get_all:[12,0,1,""],get_texts:[12,0,1,""],merge_wgts:[12,0,1,""],numericalizer:[12,0,1,""]},"pythainlp.ulmfit.utils.ThaiTokenizer":{proc_all:[12,3,1,""],proc_all_mp:[12,3,1,""],proc_text:[12,2,1,""],replace_rep:[12,3,1,""],sub_br:[12,2,1,""],tokenize:[12,2,1,""]},"pythainlp.word_vector.thai2vec":{about:[13,0,1,""],doesnt_match:[13,0,1,""],get_model:[13,0,1,""],most_similar_cosmul:[13,0,1,""],sentence_vectorizer:[13,0,1,""],similarity:[13,0,1,""]}},objnames:{"0":["py","function","Python function"],"1":["py","class","Python class"],"2":["py","method","Python method"],"3":["py","staticmethod","Python static method"]},objtypes:{"0":"py:function","1":"py:class","2":"py:method","3":"py:staticmethod"},terms:{"1cnj6o3b1jwahwvsmww3oqw7f8x2ka7sp":15,"\u0e01":[3,11,15],"\u0e01\u0e04":[11,15],"\u0e01\u0e20\u0e32\u0e29\u0e32\u0e44\u0e17\u0e22\u0e20\u0e32\u0e29\u0e32\u0e1a":[11,15],"\u0e01\u0e25\u0e21":15,"\u0e01\u0e32\u0e23\u0e43\u0e0a":13,"\u0e04":[11,12,15],"\u0e04\u0e19":[11,15],"\u0e04\u0e19\u0e44\u0e17\u0e22":[11,15],"\u0e04\u0e23":[11,15],"\u0e07":6,"\u0e07\u0e32\u0e19":13,"\u0e08":6,"\u0e13":[11,15],"\u0e13\u0e19\u0e30\u0e04\u0e23":[11,15],"\u0e14":[1,11,15],"\u0e15":2,"\u0e15\u0e32":15,"\u0e15\u0e32\u0e01":15,"\u0e17":11,"\u0e17\u0e14\u0e2a\u0e2d\u0e1a":3,"\u0e17\u0e14\u0e2a\u0e2d\u0e1a\u0e23\u0e30\u0e1a\u0e1a\u0e40\u0e27\u0e25\u0e32":3,"\u0e19":[3,11,15],"\u0e19\u0e04\u0e19\u0e44\u0e17\u0e22\u0e23":[11,15],"\u0e19\u0e17":3,"\u0e19\u0e19":6,"\u0e19\u0e30":[11,15],"\u0e1a":[11,15],"\u0e1a\u0e42\u0e2d\u0e40\u0e04\u0e1a":[11,15],"\u0e1c\u0e21":[11,15],"\u0e1c\u0e21\u0e23":[11,15],"\u0e1e\u0e27\u0e01":[11,15],"\u0e1e\u0e27\u0e01\u0e40\u0e23\u0e32":[11,15],"\u0e1e\u0e27\u0e01\u0e40\u0e23\u0e32\u0e40\u0e1b":[11,15],"\u0e20\u0e32\u0e29\u0e32":[11,15],"\u0e20\u0e32\u0e29\u0e32\u0e44\u0e17\u0e22":[11,15],"\u0e22":3,"\u0e22\u0e21\u0e32\u0e01":6,"\u0e22\u0e22\u0e22\u0e22\u0e22\u0e22":12,"\u0e23":[11,15],"\u0e23\u0e30\u0e1a\u0e1a":3,"\u0e25\u0e21":15,"\u0e25\u0e32\u0e04\u0e21":2,"\u0e27":[1,3,6],"\u0e2a\u0e23":11,"\u0e2b\u0e21":1,"\u0e2d":11,"\u0e2d\u0e14":12,"\u0e2d\u0e32\u0e01\u0e32\u0e28\u0e14":6,"\u0e2d\u0e32\u0e23\u0e21\u0e13":6,"\u0e32\u0e07\u0e08\u0e32\u0e01":11,"\u0e32\u0e19":[11,15],"\u0e32\u0e19\u0e40\u0e01":[11,15],"\u0e40\u0e01":[11,15],"\u0e40\u0e04":[11,15],"\u0e40\u0e1b":[1,11,15],"\u0e40\u0e23\u0e32":[11,15],"\u0e40\u0e27\u0e25\u0e32":3,"\u0e40\u0e2a":6,"\u0e41\u0e21\u0e27":11,"\u0e41\u0e21\u0e27\u0e14":11,"\u0e42\u0e2d":[11,15],"\u0e42\u0e2d\u0e40\u0e04":[11,15],"\u0e44\u0e01":1,"\u0e44\u0e17\u0e22":[11,15],"\u0e51":4,"\u0e52":4,"\u0e53":4,"boolean":3,"class":[3,5,12],"default":12,"function":[1,4,10,11,13],"import":[1,3,6,11,12,15],"int":[9,12],"new":[11,15],"public":13,"return":[0,1,2,3,4,5,6,7,8,10,11,12,13],"static":12,"tkrep6\u0e22":12,"true":[3,11],"try":16,EOS:15,For:[11,15,16],POS:10,The:[0,1,2,3,4,5,6,7,8,9,10,11,12,13,15],With:15,_pad_:12,_unk_:12,about:[12,13],abov:12,academ:10,accord:5,actual:11,add:12,algorithm:[8,11],all:[11,12],along:11,alphabet:[1,5],alreadi:8,amount_numb:4,analysi:6,annot:10,apach:14,archiv:16,arguabl:15,arrai:12,artagg:10,articl:10,asian:15,automat:11,avaiabl:12,base:11,basi:15,becaus:15,begin:12,being:11,best:15,between:[5,11,13],biggest:15,boil:15,bool:11,both:15,can:[15,16],captur:15,challeng:15,chang:14,charact:[4,11,12],check:[8,11],check_al:11,chines:15,choos:[5,11],chunk:11,classif:15,classifi:15,closest:8,cluster:11,cmtr:3,colab:15,collat:14,collect:12,column:12,com:[11,15,16],comma:15,common:11,connector:15,contain:[1,4,10,11,13],context:15,convert:2,corpu:10,correct:0,correctli:8,cosin:13,counter:12,cp36:16,cp36m:16,cpu:12,creat:[11,15],create_custom_dict_tri:11,current:[2,12],custom:11,custom_dict_sourc:11,custom_dict_tri:11,data:[0,1,5,11,13],data_dict:11,datafram:12,dataset:12,date:[3,14],decid:15,deep:[5,11],deepcut:11,degre:15,depend:10,desir:11,dev:16,develop:16,dict:[11,12,15],dict_word_token:11,dictionari:11,differ:[10,11,15],dim:13,document:12,document_vector:12,doe:[11,15],doesnt_match:13,doubl:12,down:15,download:13,drive:15,each:12,easi:15,edu:16,element:12,em_sz:12,embed:12,end:[11,15],engin:[5,6,8,9,10,11,12,15],english:[0,5],entir:12,entiti:3,etc:15,even:15,exampl:[1,3,6,11,12,15,16],exist:8,express:15,extract:12,extrem:15,face:15,fals:[3,11,13],field:15,file:11,find:8,first:[11,12,13],first_word:12,fix:0,follow:[12,15],format:11,found:11,french:15,freq:12,frequenc:[9,12],from:[1,2,3,6,11,12,15,16],gensim:13,get:[3,12,14],get_al:12,get_model:13,get_ner:3,get_text:12,github:[11,16],given:8,gohlk:16,googl:15,habit:15,hard:15,has:[11,15],have:15,help:[2,11],histor:15,how:5,html:11,http:[11,15,16],hunspel:8,ibm:11,icu:[11,15],includ:[12,15],incorrect:0,index:[11,14],indic:15,inform:11,input:[0,11,13],instal:[13,14],institut:5,integ:12,internaiton:5,issu:15,isthai:11,iter:12,ito:12,itos_cl:12,itos_pr:12,japanes:15,jsbr:3,known:15,label:[10,12],lack:15,languag:[0,14,15],latest:11,learn:5,less:[5,15],letter:15,lexto:11,lfd:16,librari:[14,15],licens:14,like:15,line:11,linux:8,list:[1,3,8,10,11,12,13,15],list_text:10,listdata:13,listword:11,lk82:7,longest:11,lot:15,maintain:14,make:[13,15],marisa:[11,16],marisa_tri:16,mark:[11,15],match:11,max_vocab:12,maximum:[11,12],merg:12,merge_wgt:12,meth:12,min_freq:12,minimum:12,model:[12,13],modul:14,month:2,more:[5,11,12],most:15,most_similar_cosmul:13,multi:15,multipl:[11,12],name:3,nativ:15,natur:[14,15],ncmn:3,ncnm:3,need:0,neg:[6,13],ner:14,network:11,neural:11,newlin:11,newmm:[9,11,12,15],nlp:[14,15],none:12,norvig:8,note:[14,16],now:2,num_to_text:4,num_to_thai_num:4,number:[9,12,14],numeric:12,numpi:[12,13],numtoword:4,obtain:16,often:15,old:6,one:5,option:11,orchid:10,other:15,out:[2,4,5,15],output:[2,11],overcom:15,packag:[14,15],pad:12,pad_tok:12,page:14,panda:12,parallel:10,param:[9,11,12],paramet:[0,1,3,4,5,6,7,8,10,11,12,13],part:10,path:11,perceptron:10,period:15,peter:8,phonet:5,phrase:11,pip:16,pois:15,pomm:15,pos:6,pos_tag:10,posit:[12,13],postag:3,potato:15,pre:13,pretrain:12,problem:15,proc_al:12,proc_all_mp:12,proc_text:12,process:[4,12,14,15],procss:12,prong:15,proport:11,provid:11,pud:10,punc:3,put:[5,10],pyicu:5,pylexto:[11,15],pythainlp:[15,16],python:[14,15],pythonlib:16,pytorch:12,quick:15,rather:11,rdr:10,read:5,readthedoc:11,recent:15,recogn:11,recognit:3,recommend:12,refer:[4,14],regard:10,remov:12,repetit:12,replac:12,replace_rep:12,requir:13,research:15,roman:[6,14],root:15,royal:5,royin:5,rule:[7,11,15],run:12,save:12,search:14,second:[11,12,13],second_word:12,see:11,seem:15,segment:[11,15],select:12,sens:15,sent:9,sent_token:11,sentenc:[11,12,15],sentence_vector:13,sentiment:14,set:15,should:[5,8,15],similar:[13,15],simpli:5,size:12,smallest:11,softwar:14,sort:1,soundex:14,sourc:[0,1,2,3,4,5,6,7,8,9,10,11,12,13],space:[11,12,15],special:12,speech:10,spell:[2,4,5,14],split:11,stabl:16,standard:5,start:14,stoi:12,str:[0,4,5,6,7,8,9,10,11,12,13],string:[3,10,11,12],structur:11,sub_br:12,subword:15,subword_token:11,summar:14,summarize_text:9,system:15,tag:[12,14],tagger:10,take:10,task:15,tcc:11,team:14,techniqu:15,term:15,terr:15,text:[0,1,2,3,4,5,6,8,9,10,11,12,15],text_to_num:4,texttoeng:0,texttothai:0,thai2rom:5,thai2vec:13,thai:[0,1,2,3,4,5,6,7,9,10,11,14,15],thai_num_to_num:4,thai_num_to_text:4,thainer:3,thaitoken:12,than:12,them:15,thi:[11,15],time:3,tkrep:12,tok:12,tok_engin:12,token:[9,10,12,14,15],train:[13,15],translat:15,treebank:10,trie:[11,16],tupl:[11,12],turn:5,two:[13,15],uci:16,udom83:7,ulmfit:[6,14],understand:15,unigram:10,uniqu:11,unit:11,univers:[4,10],unk_tok:12,unk_token:12,unknown:12,use:[12,13,15],use_mean:13,used:[10,11],uses:[8,11,15],using:12,util:12,vact:3,valu:11,vector:[12,13,15],veer66:11,version:16,vocab:12,vocabulari:12,vocaburari:11,weight:12,wgt:12,wheel:16,when:[11,15],where:[11,12,15],which:[8,10,11],white:11,whitespac:11,whl:16,win32:16,window:16,word1:13,word2:13,word:[4,7,8,9,10,11,12,13,15],word_token:[11,15],word_vector:14,wordcutpi:[11,15],would:15,write:15,www:16,xbo:12,year:2,yet:11,you:[11,12,15],zip:16},titles:["pythainlp.change","pythainlp.collation","pythainlp.date","pythainlp.ner","pythainlp.number","pythainlp.romanization","pythainlp.sentiment","pythainlp.soundex","pythainlp.spell","pythainlp.summarize","pythainlp.tag","pythainlp.tokenize","pythainlp.ulmfit","pythainlp.word_vector","PyThaiNLP documentation","Getting Started","Installation"],titleterms:{chang:0,collat:1,date:2,depend:13,document:14,get:15,indic:14,instal:16,modul:[0,3,9,12,13],ner:3,number:4,pythainlp:[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14],roman:5,sentiment:6,soundex:7,spell:8,start:15,summar:9,tabl:14,tag:10,token:11,ulmfit:12,word_vector:13}}) \ No newline at end of file diff --git a/docs/api/change.rst b/docs/api/change.rst new file mode 100644 index 000000000..e5b770df2 --- /dev/null +++ b/docs/api/change.rst @@ -0,0 +1,11 @@ +.. currentmodule:: pythainlp.change + +pythainlp.change +==================================== +The :class:`change` is fix incorrect input language correction. + +Modules +------- + +.. autofunction:: texttothai +.. autofunction:: texttoeng \ No newline at end of file diff --git a/docs/api/collation.rst b/docs/api/collation.rst new file mode 100644 index 000000000..a889fbea1 --- /dev/null +++ b/docs/api/collation.rst @@ -0,0 +1,8 @@ +.. currentmodule:: pythainlp.collation + + +pythainlp.collation +===================================== +The :class:`pythainlp.collation` contains a function that sorts Thai text alphabetically + +.. autofunction:: collation diff --git a/docs/api/date.rst b/docs/api/date.rst new file mode 100644 index 000000000..ec5c055d0 --- /dev/null +++ b/docs/api/date.rst @@ -0,0 +1,8 @@ +.. currentmodule:: pythainlp.date + + +pythainlp.date +===================================== +The :class:`pythainlp.date` helps output dates, as spelled out in Thai. + +.. autofunction:: now diff --git a/docs/api/ner.rst b/docs/api/ner.rst new file mode 100644 index 000000000..8810ad786 --- /dev/null +++ b/docs/api/ner.rst @@ -0,0 +1,12 @@ +.. currentmodule:: pythainlp.ner + +pythainlp.ner +==================================== +The :class:`pythainlp.ner` is named entity recognition for thai. + + +Modules +------- + +.. autoclass:: thainer + :members: get_ner diff --git a/docs/api/number.rst b/docs/api/number.rst new file mode 100644 index 000000000..459cd2b1f --- /dev/null +++ b/docs/api/number.rst @@ -0,0 +1,15 @@ +.. currentmodule:: pythainlp.number + +pythainlp.number +===================================== +The :class:`pythainlp.number` contains functions for processing thai numbers and thai words that refers to numbers. + +.. autofunction:: thai_num_to_num +.. autofunction:: thai_num_to_text +.. autofunction:: num_to_thai_num +.. autofunction:: num_to_text +.. autofunction:: text_to_num +.. autofunction:: numtowords + + + diff --git a/docs/api/romanization.rst b/docs/api/romanization.rst new file mode 100644 index 000000000..74e45f9b2 --- /dev/null +++ b/docs/api/romanization.rst @@ -0,0 +1,10 @@ +.. currentmodule:: pythainlp.romanization + +pythainlp.romanization +==================================== +The :class:`romanization.romanization` turns thai text into a romanized one (put simply, spelled with English). + +.. autofunction:: romanization +.. currentmodule:: pythainlp.romanization.thai2rom +.. autoclass:: thai2rom + :members: romanization diff --git a/docs/api/sentiment.rst b/docs/api/sentiment.rst new file mode 100644 index 000000000..10d7ce603 --- /dev/null +++ b/docs/api/sentiment.rst @@ -0,0 +1,7 @@ +.. currentmodule:: pythainlp.sentiment + +pythainlp.sentiment +==================================== +The :class:`romanization.sentiment` is sentiment analysis. + +.. autofunction:: sentiment \ No newline at end of file diff --git a/docs/api/soundex.rst b/docs/api/soundex.rst new file mode 100644 index 000000000..c61d7266f --- /dev/null +++ b/docs/api/soundex.rst @@ -0,0 +1,8 @@ +.. currentmodule:: pythainlp.soundex + +pythainlp.soundex +==================================== +The :class:`pythainlp.soundex` is soundex for thai. + +.. autofunction:: LK82 +.. autofunction:: Udom83 \ No newline at end of file diff --git a/docs/api/spell.rst b/docs/api/spell.rst new file mode 100644 index 000000000..50b379cc6 --- /dev/null +++ b/docs/api/spell.rst @@ -0,0 +1,7 @@ +.. currentmodule:: pythainlp.spell + +pythainlp.spell +===================================== +The :class:`pythainlp.spell` finds the closest correctly spelled word to the given text. + +.. autofunction:: spell diff --git a/docs/api/summarize.rst b/docs/api/summarize.rst new file mode 100644 index 000000000..d7df3988a --- /dev/null +++ b/docs/api/summarize.rst @@ -0,0 +1,10 @@ +.. currentmodule:: pythainlp.summarize + +pythainlp.summarize +==================================== +The :class:`summarize` is thai text summarize. + +Modules +------- + +.. autofunction:: summarize_text diff --git a/docs/api/tag.rst b/docs/api/tag.rst new file mode 100644 index 000000000..69315fc36 --- /dev/null +++ b/docs/api/tag.rst @@ -0,0 +1,9 @@ +.. currentmodule:: pythainlp.tag + + +pythainlp.tag +===================================== +The :class:`pythainlp.tag` contains functions that are used to tag different parts of a text. + +.. autofunction:: pos_tag + diff --git a/docs/api/tokenizer.rst b/docs/api/tokenizer.rst new file mode 100644 index 000000000..c25af5faa --- /dev/null +++ b/docs/api/tokenizer.rst @@ -0,0 +1,15 @@ +.. currentmodule:: pythainlp.tokenize +.. _tokenize-doc: + + + +pythainlp.tokenize +===================================== +The :class:`pythainlp.tokenize` contains multiple functions for tokenizing a chunk of Thai text into desirable units. + +.. autofunction:: word_tokenize +.. autofunction:: dict_word_tokenize +.. autofunction:: subword_tokenize +.. autofunction:: sent_tokenize +.. autofunction:: isthai +.. autofunction:: create_custom_dict_trie diff --git a/docs/api/ulmfit.rst b/docs/api/ulmfit.rst new file mode 100644 index 000000000..400a23899 --- /dev/null +++ b/docs/api/ulmfit.rst @@ -0,0 +1,20 @@ +.. currentmodule:: pythainlp.ulmfit.utils + + + +pythainlp.ulmfit +==================================== +The :class:`ulmfit.utils` is utils for ULMFit model. + +Modules +------- + + +.. autofunction:: get_texts +.. autofunction:: get_all +.. autofunction:: numericalizer +.. autofunction:: merge_wgts +.. autofunction:: document_vector +.. autofunction:: about +.. autoclass:: ThaiTokenizer + :members: sub_br,tokenize,replace_rep,proc_text,proc_all,proc_all_mp diff --git a/docs/api/word_vector.rst b/docs/api/word_vector.rst new file mode 100644 index 000000000..bfa44a43c --- /dev/null +++ b/docs/api/word_vector.rst @@ -0,0 +1,21 @@ +.. currentmodule:: pythainlp.word_vector.thai2vec + + + +pythainlp.word_vector +==================================== +The :class:`word_vector.thai2vec` contains functions that makes use of a pre-trained vector public data. + +Dependencies +------------ +Installation of `numpy` and `gensim` is required. + +Modules +------- + +.. autofunction:: get_model +.. autofunction:: most_similar_cosmul +.. autofunction:: doesnt_match +.. autofunction:: similarity +.. autofunction:: sentence_vectorizer +.. autofunction:: about diff --git a/docs/conf.py b/docs/conf.py new file mode 100644 index 000000000..c3f8194b8 --- /dev/null +++ b/docs/conf.py @@ -0,0 +1,188 @@ +# -*- coding: utf-8 -*- +# +# Configuration file for the Sphinx documentation builder. +# +# This file does only contain a selection of the most common options. For a +# full list see the documentation: +# http://www.sphinx-doc.org/en/master/config + +# -- Path setup -------------------------------------------------------------- + +# If extensions (or modules to document with autodoc) are in another directory, +# add these directories to sys.path here. If the directory is relative to the +# documentation root, use os.path.abspath to make it absolute, like shown here. +# +# import os +# import sys +# sys.path.insert(0, os.path.abspath('.')) +from datetime import datetime +import sys, os + +# -- Project information ----------------------------------------------------- + +project = 'PyThaiNLP' +copyright = '2018, pythainlp_builders' +author = 'pythainlp_builders' + +curyear = datetime.today().year +copyright = u'2017-%s, %s (Apache Software License 2.0)' % (curyear, project) + +# The short X.Y version +version = '' +# The full version, including alpha/beta/rc tags +release = '1.7' + + +# -- General configuration --------------------------------------------------- + +# If your documentation needs a minimal Sphinx version, state it here. +# +# needs_sphinx = '1.0' + +# Add any Sphinx extension module names here, as strings. They can be +# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom +# ones. +extensions = [ + 'sphinx.ext.autodoc', + 'sphinx.ext.doctest', + 'sphinx.ext.intersphinx', + 'sphinx.ext.todo', + 'sphinx.ext.coverage', + 'sphinx.ext.mathjax', + 'sphinx.ext.ifconfig', + 'sphinx.ext.viewcode', +] + +# Add any paths that contain templates here, relative to this directory. +templates_path = ['_templates'] + +# The suffix(es) of source filenames. +# You can specify multiple suffix as a list of string: +# +# source_suffix = ['.rst', '.md'] +source_suffix = '.rst' + +# The master toctree document. +master_doc = 'index' + +# The language for content autogenerated by Sphinx. Refer to documentation +# for a list of supported languages. +# +# This is also used if you do content translation via gettext catalogs. +# Usually you set "language" from the command line for these cases. +language = None + +# List of patterns, relative to source directory, that match files and +# directories to ignore when looking for source files. +# This pattern also affects html_static_path and html_extra_path . +exclude_patterns = ['_build', 'Thumbs.db', '.DS_Store'] + +# The name of the Pygments (syntax highlighting) style to use. +pygments_style = 'sphinx' + + +# -- Options for HTML output ------------------------------------------------- + +# The theme to use for HTML and HTML Help pages. See the documentation for +# a list of builtin themes. +# +html_theme = 'sphinx_rtd_theme' + +# Theme options are theme-specific and customize the look and feel of a theme +# further. For a list of options available for each theme, see the +# documentation. +# +# html_theme_options = {} + +# Add any paths that contain custom static files (such as style sheets) here, +# relative to this directory. They are copied after the builtin static files, +# so a file named "default.css" will overwrite the builtin "default.css". +html_static_path = ['_static'] + +# Custom sidebar templates, must be a dictionary that maps document names +# to template names. +# +# The default sidebars (for documents that don't match any pattern) are +# defined by theme itself. Builtin themes are using these templates by +# default: ``['localtoc.html', 'relations.html', 'sourcelink.html', +# 'searchbox.html']``. +# +# html_sidebars = {} + + +# -- Options for HTMLHelp output --------------------------------------------- + +# Output file base name for HTML help builder. +htmlhelp_basename = 'pythainlpdoc' + + +# -- Options for LaTeX output ------------------------------------------------ + +latex_elements = { + # The paper size ('letterpaper' or 'a4paper'). + # + # 'papersize': 'letterpaper', + + # The font size ('10pt', '11pt' or '12pt'). + # + # 'pointsize': '10pt', + + # Additional stuff for the LaTeX preamble. + # + # 'preamble': '', + + # Latex figure (float) alignment + # + # 'figure_align': 'htbp', +} + +# Grouping the document tree into LaTeX files. List of tuples +# (source start file, target name, title, +# author, documentclass [howto, manual, or own class]). +latex_documents = [ + (master_doc, 'pythainlp.tex', 'pythainlp Documentation', + 'pythainlp\\_builders', 'manual'), +] + + +# -- Options for manual page output ------------------------------------------ + +# One entry per manual page. List of tuples +# (source start file, name, description, authors, manual section). +man_pages = [ + (master_doc, 'pythainlp', 'pythainlp Documentation', + [author], 1) +] + + +# -- Options for Texinfo output ---------------------------------------------- + +# Grouping the document tree into Texinfo files. List of tuples +# (source start file, target name, title, author, +# dir menu entry, description, category) +texinfo_documents = [ + (master_doc, 'pythainlp', 'pythainlp Documentation', + author, 'pythainlp', 'One line description of project.', + 'Miscellaneous'), +] + + +# -- Extension configuration ------------------------------------------------- + +# -- Options for intersphinx extension --------------------------------------- + +# Example configuration for intersphinx: refer to the Python standard library. +intersphinx_mapping = {'python': ('https://docs.python.org/', None), + 'NLTK': ('http://nltk.org', None), + } + +# -- Options for todo extension ---------------------------------------------- + +# If true, `todo` and `todoList` produce output, else they produce nothing. +todo_include_todos = True + +# -- Markdown + +# from recommonmark.parser import CommonMarkParser +# source_parsers = {'.md': CommonMarkParser} +# source_suffix = ['.rst', '.md'] diff --git a/docs/index.md b/docs/index.md deleted file mode 100644 index 6e6ed0d66..000000000 --- a/docs/index.md +++ /dev/null @@ -1,21 +0,0 @@ -# PyThaiNLP - -Thai natural language processing in Python. - -PyThaiNLP is python module like nltk , but It's working with thai language. - -### Project status - -Developing - -### Stable Version - -1.4 - -### Dev Version - -1.5 - -### License - -Apache Software License 2.0 \ No newline at end of file diff --git a/docs/index.rst b/docs/index.rst new file mode 100644 index 000000000..c5e2d89cc --- /dev/null +++ b/docs/index.rst @@ -0,0 +1,36 @@ +.. pythainlp documentation master file, created by + sphinx-quickstart on Sat Jun 23 15:23:30 2018. + You can adapt this file completely to your liking, but it should at least + contain the root `toctree` directive. + +PyThaiNLP documentation +===================================== +PyThaiNLP is a Python library for natural language processing (NLP) of Thai language. + + +.. toctree:: + :glob: + :maxdepth: 1 + :caption: Notes + + notes/* + +.. toctree:: + :glob: + :maxdepth: 1 + :caption: Package reference: + + api/* + + + +Indices and tables +================== + +* :ref:`genindex` +* :ref:`modindex` +* :ref:`search` + +Apache Software License 2.0 + +Maintained by the PyThaiNLP team. diff --git a/docs/notes/getting_started.rst b/docs/notes/getting_started.rst new file mode 100644 index 000000000..2fc1285c6 --- /dev/null +++ b/docs/notes/getting_started.rst @@ -0,0 +1,27 @@ +Getting Started +===================================== +PyThaiNLP is a Python library for natural language processing (NLP) of Thai language. With this package, you can NLP tasks such as classify texts and tokenize words. + +**Tokenization Example**:: + + from pythainlp.tokenize import word_tokenize + text='ผมรักคุณนะครับโอเคบ่พวกเราเป็นคนไทยรักภาษาไทยภาษาบ้านเกิด' + a=word_tokenize(text,engine='icu') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอ', 'เค', 'บ่', 'พวก', 'เรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้าน', 'เกิด'] + b=word_tokenize(text,engine='dict') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + c=word_tokenize(text,engine='mm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + d=word_tokenize(text,engine='pylexto') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + e=word_tokenize(text,engine='newmm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] + g=word_tokenize(text,engine='wordcutpy') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้านเกิ + +Thai has historically faced a lot of NLP challenges. A quick list of them include as follows: + +#. **Start-end of sentence marking** - This is arguably the biggest problem for the field of Thai NLP. The lack of end of sentence marking (EOS) makes it hard for researchers to create training sets, the basis of most research in this field. The root of the problem is two-pronged. In terms of writing system, Thai uses space to indicate both commas and periods. No letter indicates an end of a sentence. In terms of language use, Thais have a habit of starting their sentences with connector terms such as 'because', 'but', 'following', etc, making it often hard even for natives to decide where the end of sentence should be when translating. + +#. **Word segmentation** - Thai does not use space and word segmentation is not easy. It boils down to understanding the context and ruling out words that do not make sense. This is a similar issue that other Asian languages such as Japanese and Chinese face in different degrees. For languages with space, a similar but less extreme problem would be multi-word expressions, like the French word for potato — 'pomme de terre'. In Thai, the best known example is "ตา-กลม" and "ตาก-ลม". As of recent, new techniques that capture words, subwords, and letters in vectors seem poised to overcome to issue. + +**Example for classification** + +https://colab.research.google.com/drive/1cnJ6O3b1jwaHwvsMWW3oQw7f8X2Ka7Sp + + + diff --git a/docs/notes/installation.rst b/docs/notes/installation.rst new file mode 100644 index 000000000..49b251b65 --- /dev/null +++ b/docs/notes/installation.rst @@ -0,0 +1,11 @@ +Installation +===================================== +For stable version, try:: + + pip install pythainlp + +For developer version, try:: + + pip install https://github.com/PyThaiNLP/pythainlp/archive/dev.zip + +Note for Windows: marisa-trie wheels can be obtained from https://www.lfd.uci.edu/~gohlke/pythonlibs/#marisa-trie , then install it with pip, for example: `pip install marisa_trie‑0.7.5‑cp36‑cp36m‑win32.whl` \ No newline at end of file diff --git a/docs/pythainlp-1-6-eng.md b/docs/pythainlp-1-6-eng.md new file mode 100644 index 000000000..df1c9f16e --- /dev/null +++ b/docs/pythainlp-1-6-eng.md @@ -0,0 +1,502 @@ +# User manual PyThaiNLP 1.6 + +[TOC] + +## API + +### tokenize + +#### word_tokenize + +word_tokenize is thai word segmatation. + +```python +from pythainlp.tokenize import word_tokenize +word_tokenize(text,engine) +``` +**text** refers to an input text string in Thai. + +**engine** refers to a thai word segmentation system; There are 6 systems to choose from. + +1. newmm (default) - Maximum Matching algorithm for Thai word segmatation. Developed by Korakot Chaovavanich (https://www.facebook.com/groups/408004796247683/permalink/431283740586455/) +2. icu - pyicu has a very poor performance. +3. dict - dictionary-based tokenizer. It returns False if the message can not be wrapped. +4. longest-matching - using Longest matching algorithm for Thai word segmentation. +5. mm - Maximum Matching algorithm for Thai word segmentation. +6. pylexto - LexTo. +7. deepcut - Deep Learning based Thai word segmentation (https://github.com/rkcosmos/deepcut) +8. wordcutpy - use wordcutpy (https://github.com/veer66/wordcutpy) for Thai word segmentation. + + +Output: ''list'' ex. ['แมว','กิน'] + +**Example** + + +```python +from pythainlp.tokenize import word_tokenize +text='ผมรักคุณนะครับโอเคบ่พวกเราเป็นคนไทยรักภาษาไทยภาษาบ้านเกิด' +a=word_tokenize(text,engine='icu') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอ', 'เค', 'บ่', 'พวก', 'เรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้าน', 'เกิด'] +b=word_tokenize(text,engine='dict') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] +c=word_tokenize(text,engine='mm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] +d=word_tokenize(text,engine='pylexto') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] +e=word_tokenize(text,engine='newmm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] +g=word_tokenize(text,engine='wordcutpy') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้านเกิด'] +``` + +#### dict_word_tokenize + +```python +from pythainlp.tokenize import dict_word_tokenize +dict_word_tokenize(text,file,engine) +``` + +A command for tokenize by using user-defined information. + +text : str + +file : name file data using in tokenize. + +engine + +- newmm +- wordcutpy : using wordcutpy (https://github.com/veer66/wordcutpy) +- mm +- longest-matching + +Example https://gist.github.com/wannaphongcom/1e862583051bf0464b6ef4ed592f739c + +#### sent_tokenize + +Thai Sentence Tokenizer + +```python +sent_tokenize(text,engine='whitespace+newline') +``` + +engine : + +- whitespace - tokenizer from whitespace +- whitespace+newline - tokenizer from whitespace and newline. + +#### Thai Character Clusters (TCC) + +TCC : Mr.Jakkrit TeCho + +grammar : Wittawat Jitkrittum (https://github.com/wittawatj/jtcc/blob/master/TCC.g) + +Code : Korakot Chaovavanich + +**Example** + +```python +>>> from pythainlp.tokenize import tcc +>>> tcc.tcc('ประเทศไทย') +'ป/ระ/เท/ศ/ไท/ย' +``` + +#### Enhanced Thai Character Cluster (ETCC) + +**Example** + +```python +>>> from pythainlp.tokenize import etcc +>>> etcc.etcc('คืนความสุข') +'/คืน/ความสุข' +``` + +#### WhitespaceTokenizer + +Tokenizer by using spaces + +```python +>>> from pythainlp.tokenize import WhitespaceTokenizer +>>> WhitespaceTokenizer("ทดสอบ ตัดคำช่องว่าง") +['ทดสอบ', 'ตัดคำช่องว่าง'] +``` +#### isthai + +check + +### Thai postaggers + +```python +from pythainlp.tag import pos_tag +pos_tag(list,engine='old') +``` + +engine + +1. old is the UnigramTagger (default) +2. artagger is the RDR POS Tagger. + +### romanization + +```python +from pythainlp.romanization import romanization +romanization(str,engine='pyicu') +``` +There are 2 engines + +- pyicu +- royin + +data : + +input ''str'' + +returns ''str'' + +**Example** + +```python +from pythainlp.romanization import romanization +romanization("แมว") # 'mæw' +``` + +### keywords + +#### find_keyword + +find keywords from thai text in list. + +```python +find_keyword(list,lentext=3) +``` +lentext is minimum number of keywords words. + +return dict {words: number of keywords} + +### Spell Check + +```python +spell(word,engine='pn') +``` +engine + +- 'pn' code from Peter Norvig +- 'hunspell' using hunspell + +Before using this module, please install hunspell and hunspell-th. + +```python +from pythainlp.spell import * +a=spell("สี่เหลียม") +print(a) # ['สี่เหลี่ยม'] +``` +#### pn + +```python +correction(word) +``` + +Show word possible + +**Sample usage** + +```python +from pythainlp.spell.pn import correction +a=correction("สี่เหลียม") +print(a) # ['สี่เหลี่ยม'] +``` +### number + +```python +from pythainlp.number import * +``` +- nttn(str) - convert thai numbers to numbers. +- nttt(str) - Thai Numbers to text. +- ntnt(str) - numbers to thai numbers. +- ntt(str) - numbers to text. +- ttn(str) - text to numbers. +- numtowords(float) - Read thai numbers (Baht) input ''float'' returns 'str' + +### Sort Thai text into List + +```python +from pythainlp.collation import collation +print(collation(['ไก่','ไข่','ก','ฮา'])) # ['ก', 'ไก่', 'ไข่', 'ฮา'] +``` + +input list + +returns list + +### Get current time in Thai + +```python +from pythainlp.date import now +now() # '30 พฤษภาคม 2560 18:45:24' +``` +### Find the most frequent words. + +```python +from pythainlp.rank import rank +rank(list) +``` + +returns dict + +**Example** + +```python +>>> rank(['แมง','แมง','คน']) +Counter({'แมง': 2, 'คน': 1}) +``` + +### Incorrect input language correction + +```python +from pythainlp.change import * +``` + +- texttothai(str) - eng to thai. +- texttoeng(str) - thai to eng. + +### word_vector + +```python +from pythainlp.word_vector import thai2vec +``` + +word_vector is word vector in PyThaiNLP + +It's work using thai2vec (https://github.com/cstorm125/thai2vec) + +thai2vec developed by Charin Polpanumas + +#### thai2vec + +requirements + +- gensim +- numpy + +##### API + +- get_model() - get gensim model +- most_similar_cosmul(positive,negative) +- doesnt_match(listdata) +- similarity(word1,word2) +- sentence_vectorizer(ss,dim=300,use_mean=False) +- about() + +### Thai Soundex + +credit Korakot Chaovavanich (from https://gist.github.com/korakot/0b772e09340cac2f493868da035597e8) + +- LK82 +- Udom83 + +**Example** + +```python +>>> from pythainlp.soundex import LK82,Udom83 +>>> print(LK82('รถ')) +ร3000 +>>> print(LK82('รด')) +ร3000 +>>> print(LK82('จัน')) +จ4000 +>>> print(LK82('จันทร์')) +จ4000 +>>> print(Udom83('รถ')) +ร800000 +``` + +### Thai meta sound + +``` +Snae & Brückner. (2009). Novel Phonetic Name Matching Algorithm with a Statistical Ontology for Analysing Names Given in Accordance with Thai Astrology. Retrieved from https://pdfs.semanticscholar.org/3983/963e87ddc6dfdbb291099aa3927a0e3e4ea6.pdf +``` + +**Example** + +```python +>>> from pythainlp.MetaSound import * +>>> MetaSound('คน') +'15' +``` + +### Thai sentiment analysis + +using data from [https://github.com/wannaphongcom/lexicon-thai/tree/master/ข้อความ/](https://github.com/wannaphongcom/lexicon-thai/tree/master/ข้อความ/) + +```python +from pythainlp.sentiment import sentiment +sentiment(str) +``` + +input str returns pos , neg or neutral + +### Util + +using + +```python +from pythainlp.util import * +``` + +#### ngrams + +for building ngrams + +```python +ngrams(token,num) +``` + +- token - list +- num - ngrams + +#### bigrams + +for building bigrams + +```python +bigrams(token) +``` + +- token - list + +#### trigram + +for building trigram + +```python +trigram(token) +``` + +- token - list + +#### normalize + +fix thai text + +```python +normalize(text) +``` + +**Example** + +```python +>>> print(normalize("เเปลก")=="แปลก") # เ เ ป ล ก กับ แปลก +True +``` + +### Corpus + +#### Thai stopword + +```python +from pythainlp.corpus import stopwords +stopwords = stopwords.words('thai') +``` + +#### Thai country name + +```python +from pythainlp.corpus import country +country.get_data() +``` + +#### Tone in Thai + +```python +from pythainlp.corpus import tone +tone.get_data() +``` + +#### Consonant in thai + +```python +from pythainlp.corpus import alphabet +alphabet.get_data() +``` + +#### Word list in thai + +```python +from pythainlp.corpus.thaiword import get_data # old data +get_data() +from pythainlp.corpus.newthaiword import get_data # new data +get_data() +``` + +#### ConceptNet + +Thai tool for ConceptNet. + +**find edges** + +```python +edges(word,lang='th') +``` + +return dict + +#### Thai WordNet + +import + +```python +from pythainlp.corpus import wordnet +``` + +**Use** + +It's like nltk. + +- wordnet.synsets(word) +- wordnet.synset(name_synsets) +- wordnet.all_lemma_names(pos=None, lang="tha") +- wordnet.all_synsets(pos=None) +- wordnet.langs() +- wordnet.lemmas(word,pos=None,lang="tha") +- wordnet.lemma(name_synsets) +- wordnet.lemma_from_key(key) +- wordnet.path_similarity(synsets1,synsets2) +- wordnet.lch_similarity(synsets1,synsets2) +- wordnet.wup_similarity(synsets1,synsets2) +- wordnet.morphy(form, pos=None) +- wordnet.custom_lemmas(tab_file, lang) + +**Example** + +```python +>>> from pythainlp.corpus import wordnet +>>> print(wordnet.synsets('หนึ่ง')) +[Synset('one.s.05'), Synset('one.s.04'), Synset('one.s.01'), Synset('one.n.01')] +>>> print(wordnet.synsets('หนึ่ง')[0].lemma_names('tha')) +[] +>>> print(wordnet.synset('one.s.05')) +Synset('one.s.05') +>>> print(wordnet.synset('spy.n.01').lemmas()) +[Lemma('spy.n.01.spy'), Lemma('spy.n.01.undercover_agent')] +>>> print(wordnet.synset('spy.n.01').lemma_names('tha')) +['สปาย', 'สายลับ'] +``` + +#### TNC + +Tool for Thai National Corpus (http://www.arts.chula.ac.th/~ling/TNC/index.php) + +##### word_frequency + +find word frequency + +```python +word_frequency(word,domain='all') +``` +domain + +- all +- imaginative +- natural-pure-science +- applied-science +- social-science +- world-affairs-history +- commerce-finance +- arts +- belief-thought +- leisure +- others \ No newline at end of file diff --git a/docs/pythainlp-1-6-thai.md b/docs/pythainlp-1-6-thai.md index 6018b32aa..c3a9ed3cd 100644 --- a/docs/pythainlp-1-6-thai.md +++ b/docs/pythainlp-1-6-thai.md @@ -22,7 +22,7 @@ pip install pythainlp วิธีที่ง่ายที่สุดคือใช้ wheel 1. [http://www.lfd.uci.edu/~gohlke/pythonlibs/#pyicu](http://www.lfd.uci.edu/~gohlke/pythonlibs/#pyicu) แล้ว download wheel ตาม python ตัวเองเช่น - ผมใช้ python x64 3.6.1 บน Windows ก็ให้ใช้ PyICU‑1.9.7‑cp36‑cp36m‑win_amd64.whl + ผมใช้ python x64 3.6.1 บน Windows ก็ให้ใช้ PyICU‑1.9.7‑cp36‑cp36m‑win_amd64.whl 2. `pip install PyICU‑1.9.7‑cp36‑cp36m‑win_amd64.whl` @@ -61,7 +61,7 @@ engine คือ ระบบตัดคำไทย ปัจจุบัน 3. dict - เป็นการตัดคำโดยใช้พจานุกรมจาก thaiword.txt ใน corpus (ความแม่นยำปานกลาง) **จะคืนค่า False หากข้อความนั้นไม่สามารถตัดคำได้** 4. longest-matching ใช้ Longest matching ในการตัดคำ 5. mm - ใช้ Maximum Matching algorithm ในการตัดคำภาษาไทย - API ชุดเก่า **อยู่ในหมวดบำรุงรักษาเท่านั้น** -6. ****pylexto ใช้ LexTo ในการตัดคำ โดยเป็น Longest matching +6. pylexto ใช้ LexTo ในการตัดคำ โดยเป็น Longest matching 7. deepcut ใช้ deepcut จาก https://github.com/rkcosmos/deepcut ในการตัดคำภาษาไทย 8. wordcutpy ใช้ wordcutpy (https://github.com/veer66/wordcutpy) ในการตัดคำ @@ -173,7 +173,7 @@ check_all สำหรับส่งคืนค่า True หรือ False รองรับ Thai Character Clusters (TCC) โดยจะแบ่งกลุ่มด้วย / -**เดติด** +**เครดิต** TCC : Mr.Jakkrit TeCho @@ -201,6 +201,22 @@ grammar : คุณ Wittawat Jitkrittum (https://github.com/wittawatj/jtcc/blob/ '/คืน/ความสุข' ``` +### tag + +เป็น Part-of-speech tagging ภาษาไทย + +```python +from pythainlp.tag import pos_tag +pos_tag(list,engine='old') +``` + +list คือ list ที่เก็บข้อความหลังผ่านการตัดคำแล้ว + +engine คือ ชุดเครื่องมือในการ postaggers มี 2 ตัวดังนี้ + +1. old เป็น UnigramTagger (ค่าเริ่มต้น) +2. artagger เป็น RDR POS Tagger ละเอียดยิ่งกว่าเดิม รองรับเฉพาะ Python 3 เท่านั้น + ### summarize เป็นระบบสรุปเอกสารภาษาไทยแบบง่าย ๆ @@ -267,22 +283,6 @@ lentext คือ จำนวนคำขั้นต่ำที่ต้อ คืนค่าออกมาเป็น dict -### tag - -เป็น Part-of-speech tagging ภาษาไทย - -```python -from pythainlp.tag import pos_tag -pos_tag(list,engine='old') -``` - -list คือ list ที่เก็บข้อความหลังผ่านการตัดคำแล้ว - -engine คือ ชุดเครื่องมือในการ postaggers มี 2 ตัวดังนี้ - -1. old เป็น UnigramTagger (ค่าเริ่มต้น) -2. artagger เป็น RDR POS Tagger ละเอียดยิ่งกว่าเดิม รองรับเฉพาะ Python 3 เท่านั้น - ### romanization ```python diff --git a/docs/pythainlp-dev-thai.md b/docs/pythainlp-dev-thai.md new file mode 100644 index 000000000..23a80aa60 --- /dev/null +++ b/docs/pythainlp-dev-thai.md @@ -0,0 +1,702 @@ +# คู่มือการใช้งาน PyThaiNLP 1.7 + +[TOC] + +Natural language processing หรือ การประมวลภาษาธรรมชาติ โมดูล PyThaiNLP เป็นโมดูลที่ถูกพัฒนาขึ้นเพื่อพัฒนาการประมวลภาษาธรรมชาติภาษาไทยในภาษา Python และ**มันฟรี (ตลอดไป) เพื่อคนไทยและชาวโลกทุกคน !** + +> เพราะโลกขับเคลื่อนต่อไปด้วยการแบ่งปัน + +รองรับ Python 2.7 และ Python 3.4 ขึ้นไปเท่านั้น + +ติดตั้งใช้คำสั่ง + +``` +pip install pythainlp +``` + +ปัจจุบันนี้ PyThaiNLP ไม่ต้องการ PyICU ในการใช้งาน API อีกแล้ว แต่หากท่านต้องการใช้ API ที่มี PyICU ให้ทำตามคำแนะนำข้างล่างนี้ + +**วิธีติดตั้งสำหรับ Windows** + +การติดตั้ง pythainlp บน windows ต้องติดตั้ง pyicu ซึ่งทำได้ยากมาก +วิธีที่ง่ายที่สุดคือใช้ wheel + +1. [http://www.lfd.uci.edu/~gohlke/pythonlibs/#pyicu](http://www.lfd.uci.edu/~gohlke/pythonlibs/#pyicu) แล้ว download wheel ตาม python ตัวเองเช่น + ผมใช้ python x64 3.6.1 บน Windows ก็ให้ใช้ PyICU‑1.9.7‑cp36‑cp36m‑win_amd64.whl + +2. `pip install PyICU‑1.9.7‑cp36‑cp36m‑win_amd64.whl` + +3. `pip install pythainlp` + +**ติดตั้งบน Mac** + +** แนะนำให้ใช้ icu 58.2 เนื่องจาก icu 59.1 มาปัญหากับ PyICU ** + +```sh +$ brew install icu4c --force +$ brew link --force icu4c +$ CFLAGS=-I/usr/local/opt/icu4c/include LDFLAGS=-L/usr/local/opt/icu4c/lib pip install pythainlp +``` + +ข้อมูลเพิ่มเติม [คลิกที่นี้](https://medium.com/data-science-cafe/install-polyglot-on-mac-3c90445abc1f#.rdfrorxjx) + +## API + +### tokenize + +#### word_tokenize + +สำหรับการตัดคำไทยนั้น ใช้ API ดังต่อไปนี้ + +```python +from pythainlp.tokenize import word_tokenize +word_tokenize(text,engine) +``` +text คือ ข้อความในรูปแบบสตริง str เท่านั้น + +engine คือ ระบบตัดคำไทย ปัจจุบันนี้ PyThaiNLP ได้พัฒนามี 6 engine ให้ใช้งานกันดังนี้ + +1. newmm - ใช้ Maximum Matching algorithm ในการตัดคำภาษาไทย โค้ดชุดใหม่ โดยใช้โค้ดคุณ Korakot Chaovavanich จาก https://www.facebook.com/groups/408004796247683/permalink/431283740586455/ มาพัฒนาต่อ (ค่าเริ่มต้น) +2. icu - engine ตัวดั้งเดิมของ PyThaiNLP (ความแม่นยำต่ำ) +3. dict - เป็นการตัดคำโดยใช้พจานุกรมจาก thaiword.txt ใน corpus (ความแม่นยำปานกลาง) **จะคืนค่า False หากข้อความนั้นไม่สามารถตัดคำได้** +4. longest-matching ใช้ Longest matching ในการตัดคำ +5. mm - ใช้ Maximum Matching algorithm ในการตัดคำภาษาไทย - API ชุดเก่า **อยู่ในหมวดบำรุงรักษาเท่านั้น** +6. pylexto ใช้ LexTo ในการตัดคำ โดยเป็น Longest matching +7. deepcut ใช้ deepcut จาก https://github.com/rkcosmos/deepcut ในการตัดคำภาษาไทย +8. wordcutpy ใช้ wordcutpy (https://github.com/veer66/wordcutpy) ในการตัดคำ + +คืนค่าเป็น ''list'' เช่น ['แมว','กิน'] + +**ตัวอย่าง** + +``` +สำหรับผู้ใช้งาน Python 2.7 ให้ทำการ encode ให้เป็น UTF-8 ก่อนใช้งานโมดูล PyThaiNLP + +เช่น text=u'ผมรักคุณนะครับโอเคบ่พวกเราเป็นคนไทยรักภาษาไทยภาษาบ้านเกิด' +``` + +**การใช้งาน** + +```python +from pythainlp.tokenize import word_tokenize +text='ผมรักคุณนะครับโอเคบ่พวกเราเป็นคนไทยรักภาษาไทยภาษาบ้านเกิด' +a=word_tokenize(text,engine='icu') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอ', 'เค', 'บ่', 'พวก', 'เรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้าน', 'เกิด'] +b=word_tokenize(text,engine='dict') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] +c=word_tokenize(text,engine='mm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] +d=word_tokenize(text,engine='pylexto') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] +e=word_tokenize(text,engine='newmm') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คนไทย', 'รัก', 'ภาษาไทย', 'ภาษา', 'บ้านเกิด'] +g=word_tokenize(text,engine='wordcutpy') # ['ผม', 'รัก', 'คุณ', 'นะ', 'ครับ', 'โอเค', 'บ่', 'พวกเรา', 'เป็น', 'คน', 'ไทย', 'รัก', 'ภาษา', 'ไทย', 'ภาษา', 'บ้านเกิด'] +``` + +#### dict_word_tokenize + +```python +from pythainlp.tokenize import dict_word_tokenize +dict_word_tokenize(text,file,engine) +``` + +เป็นคำสั่งสำหรับตัดคำโดยใช้ข้อมูลที่ผู้ใช้กำหนด + +text คือ ข้อความที่ต้องการตัดคำ + +file คือ ที่ตั้งไฟล์ที่ต้องการมาเป็นฐานข้อมูลตัดคำ + +engine คือ เครื่องมือตัดคำ + +- newmm ตัดคำด้วย newmm +- wordcutpy ใช้ wordcutpy (https://github.com/veer66/wordcutpy) ในการตัดคำ +- mm ตัดคำด้วย mm +- longest-matching ตัดคำโดยใช้ longest matching + +ตัวอย่างการใช้งาน https://gist.github.com/wannaphongcom/1e862583051bf0464b6ef4ed592f739c + +``` +สำหรับผู้ใช้งาน Python 2.7 ให้ทำการ encode ให้เป็น UTF-8 ก่อนใช้งานโมดูล PyThaiNLP + +เช่น text=u'ผมรักคุณนะครับโอเคบ่พวกเราเป็นคนไทยรักภาษาไทยภาษาบ้านเกิด' +``` + +#### sent_tokenize + +ใช้ตัดประโยคภาษาไทย + +```python +sent_tokenize(text,engine='whitespace+newline') +``` + +text คือ ข้อความในรูปแบบสตริง + +engine คือ เครื่องมือสำหรับใช้ตัดประโยค + +- whitespace ตัดประโยคจากช่องว่าง +- whitespace+newline ตัดประโยคจากช่องว่างและตัดจากการขึ้นบรรทัดใหม่ + +คืนค่า ออกมาเป็น list + +#### WhitespaceTokenizer + +ใช้ตัดคำ/ประโยคจากช่องว่างในสตริง + +```python +>>> from pythainlp.tokenize import WhitespaceTokenizer +>>> WhitespaceTokenizer("ทดสอบ ตัดคำช่องว่าง") +['ทดสอบ', 'ตัดคำช่องว่าง'] +``` + +``` +สำหรับผู้ใช้งาน Python 2.7 ให้ทำการ encode ให้เป็น UTF-8 ก่อนใช้งานโมดูล PyThaiNLP + +เช่น WhitespaceTokenizer(u"ทดสอบ ตัดคำช่องว่าง") +``` + + + +#### isthai + +ใช้เช็คข้อความว่าเป็นภาษาไทยทั้งหมดกี่ % + +```python +isthai(text,check_all=False) +``` + +text คือ ข้อความหรือ list ตัวอักษร + +check_all สำหรับส่งคืนค่า True หรือ False เช็คทุกตัวอักษร + +**การส่งคืนค่า** + +```python +{'thai':% อักษรภาษาไทย,'check_all':tuple โดยจะเป็น (ตัวอักษร,True หรือ False)} +``` + +#### Thai Character Clusters (TCC) + +รองรับ Thai Character Clusters (TCC) โดยจะแบ่งกลุ่มด้วย / + +**เครดิต** + +TCC : Mr.Jakkrit TeCho + +grammar : คุณ Wittawat Jitkrittum (https://github.com/wittawatj/jtcc/blob/master/TCC.g) + +โค้ด : คุณ Korakot Chaovavanich + +**การใช้งาน** + +```python +>>> from pythainlp.tokenize import tcc +>>> tcc.tcc('ประเทศไทย') +'ป/ระ/เท/ศ/ไท/ย' +``` + +#### Enhanced Thai Character Cluster (ETCC) + +นอกจาก TCC แล้ว PyThaiNLP 1.4 ยังรองรับ Enhanced Thai Character Cluster (ETCC) โดยแบ่งกลุ่มด้วย / + +**การใช้งาน** + +```python +>>> from pythainlp.tokenize import etcc +>>> etcc.etcc('คืนความสุข') +'/คืน/ความสุข' +``` + +### tag + +เป็น Part-of-speech tagging ภาษาไทย + +```python +from pythainlp.tag import pos_tag +pos_tag(text,engine='unigram',corpus='orchid') +``` + +list คือ list ที่เก็บข้อความหลังผ่านการตัดคำแล้ว + +engine คือ ชุดเครื่องมือในการ postaggers มี 2 ตัวดังนี้ + +1. unigram เป็น UnigramTagger (ค่าเริ่มต้น) +2. perceptron เป็น PerceptronTagger +3. artagger เป็น RDR POS Tagger ละเอียดยิ่งกว่าเดิม รองรับเฉพาะ Python 3 เท่านั้น + +corpus ที่รองรับ + +1. orchid +2. pud ใช้ข้อมูล Parallel Universal Dependencies (PUD) treebanks + +### summarize + +เป็นระบบสรุปเอกสารภาษาไทยแบบง่าย ๆ + +summarize_text(text,n,engine='frequency') + + text เป็นข้อความ + n คือ จำนวนประโยคสรุป + engine ที่รองรับ + - frequency +**การใช้งาน** + +```python +>>> from pythainlp.summarize import summarize_text +>>> summarize_text(text="อาหาร หมายถึง ของแข็งหรือของเหลว ที่กินหรือดื่มเข้าสู่ร่างกายแล้ว จะทำให้เกิดพลังงานและความร้อนยเจริญเติบโต ซ่อมแซมส่วนที่สึกหรอ ควบคุมการเปลี่ยนแปลงต่างๆ ในร่างกาย ช่วยทำให้อวัยวะต่างๆ ทำงานได้อย่างปกติ อาหารจะต้องงกาย",n=1,engine='frequency') +['อาหารจะต้องไม่มีพิษและไม่เกิดโทษต่อร่างกาย'] +``` + +### word_vector + +```python +from pythainlp.word_vector import thai2vec +``` + +word_vector เป็นระบบ word vector ใน PyThaiNLP + +ปัจจุบันนี้รองรับเฉพาะ thai2vec (https://github.com/cstorm125/thai2vec) + +thai2vec พัฒนาโดยคุณ Charin Polpanumas + +#### thai2vec + +ความต้องการโมดูล + +- gensim +- numpy + +##### API + +- get_model() - รับข้อมูล model ในรูปแบบของ gensim +- most_similar_cosmul(positive,negative) +- doesnt_match(listdata) +- similarity(word1,word2) - หาค่าความคล้ายกันระหว่าง 2 คำ โดยทั้งคู่เป็น str +- sentence_vectorizer(ss,dim=300,use_mean=False) +- about() - รายละเอียด thai2vec + + + +### keywords + +ใช้หา keywords จากข้อความภาษาไทย + +#### find_keyword + +การทำงาน หาคำที่ถูกใช้งานมากกว่าค่าขั้นต่ำที่กำหนดได้ โดยจะลบ stopword ออกไป + +```python +find_keyword(word_list,lentext=3) +``` + +word_list คือ list ของข้อความที่ผ่านการตัดคำแล้ว + +lentext คือ จำนวนคำขั้นต่ำที่ต้องการหา keyword + +คืนค่าออกมาเป็น dict + +### romanization + +```python +from pythainlp.romanization import romanization +romanization(str,engine='royin') +``` +มี 2 engine ดังนี้ + +- pyicu ส่งค่า Latin +- royin ใช้หลักเกณฑ์การถอดอักษรไทยเป็นอักษรโรมัน ฉบับราชบัณฑิตยสถาน (**หากมีข้อผิดพลาด ให้ใช้คำอ่าน เนื่องจากตัว royin ไม่มีตัวแปลงคำเป็นคำอ่าน**) + +data : + +รับค่า ''str'' ข้อความ + +คืนค่าเป็น ''str'' ข้อความ + +**ตัวอย่าง** + +```python +from pythainlp.romanization import romanization +romanization("แมว") # 'maew' +``` + +### spell + +เป็น API สำหรับเช็คคำผิดในภาษาไทย + +```python +spell(word,engine='pn') +``` + +engine ที่รองรับ + +- pn พัฒนามาจาก Peter Norvig (ค่าเริ่มต้น) +- hunspell ใช้ hunspell (ไม่รองรับ Python 2.7) + +**ตัวอย่างการใช้งาน** + +```python +from pythainlp.spell import * +a=spell("สี่เหลียม") +print(a) # ['สี่เหลี่ยม'] +``` +#### pn + +```python +correction(word) +``` + +แสดงคำที่เป็นไปได้มากที่สุด + +**ตัวอย่างการใช้งาน** + +```python +from pythainlp.spell.pn import correction +a=correction("สี่เหลียม") +print(a) # ['สี่เหลี่ยม'] +``` + +ผลลัพธ์ + +``` +สี่เหลี่ยม +``` + +### pythainlp.number + +```python +from pythainlp.number import * +``` +จัดการกับตัวเลข โดยมีดังนี้ + +- thai_num_to_num(str) - เป็นการแปลงเลขไทยสู่เลข +- thai_num_to_text(str) - เลขไทยสู่ข้อความ +- num_to_thai_num(str) - เลขสู่เลขไทย +- num_to_text(str) - เลขสู่ข้อความ +- text_to_num(str) - ข้อความสู่เลข +- numtowords(float) - อ่านจำนวนตัวเลขภาษาไทย (บาท) รับค่าเป็น ''float'' คืนค่าเป็น 'str' + +### collation + +ใช้ในการเรียงลำดับข้อมูลภาษาไทยใน List + +```python +from pythainlp.collation import collation +print(collation(['ไก่','ไข่','ก','ฮา'])) # ['ก', 'ไก่', 'ไข่', 'ฮา'] +``` + +รับ list คืนค่า list + +### date + +#### now + +รับเวลาปัจจุบันเป็นภาษาไทย + +```python +from pythainlp.date import now +now() # '30 พฤษภาคม 2560 18:45:24' +``` +### rank + +#### rank + +หาคำที่มีจำนวนการใช้งานมากที่สุด + +```python +from pythainlp.rank import rank +rank(list) +``` + +คืนค่าออกมาเป็น dict + +**ตัวอย่างการใช้งาน** + +```python +>>> rank(['แมง','แมง','คน']) +Counter({'แมง': 2, 'คน': 1}) +``` + +### change + +#### แก้ไขปัญหาการพิมพ์ลืมเปลี่ยนภาษา + +```python +from pythainlp.change import * +``` + +มีคำสั่งดังนี้ + +- texttothai(str) แปลงแป้นตัวอักษรภาษาอังกฤษเป็นภาษาไทย +- texttoeng(str) แปลงแป้นตัวอักษรภาษาไทยเป็นภาษาอังกฤษ + +คืนค่าออกมาเป็น str + +### soundex + +เดติด คุณ Korakot Chaovavanich (จาก https://gist.github.com/korakot/0b772e09340cac2f493868da035597e8) + +กฎที่รองรับในเวชั่น 1.4 + +- กฎการเข้ารหัสซาวน์เด็กซ์ของ วิชิตหล่อจีระชุณห์กุล และ เจริญ คุวินทร์พันธุ์ - LK82 +- กฎการเข้ารหัสซาวน์เด็กซ์ของ วรรณี อุดมพาณิชย์ - Udom83 + +**การใช้งาน** + +```python +>>> from pythainlp.soundex import LK82,Udom83 +>>> print(LK82('รถ')) +ร3000 +>>> print(LK82('รด')) +ร3000 +>>> print(LK82('จัน')) +จ4000 +>>> print(LK82('จันทร์')) +จ4000 +>>> print(Udom83('รถ')) +ร800000 +``` + +### Meta Sound ภาษาไทย + +``` +Snae & Brückner. (2009). Novel Phonetic Name Matching Algorithm with a Statistical Ontology for Analysing Names Given in Accordance with Thai Astrology. Retrieved from https://pdfs.semanticscholar.org/3983/963e87ddc6dfdbb291099aa3927a0e3e4ea6.pdf +``` + +**การใช้งาน** + +```python +>>> from pythainlp.MetaSound import * +>>> MetaSound('คน') +'15' +``` + +### sentiment + +เป็น Sentiment analysis ภาษาไทย ใช้ข้อมูลจาก [https://github.com/wannaphongcom/lexicon-thai/tree/master/ข้อความ/](https://github.com/wannaphongcom/lexicon-thai/tree/master/ข้อความ/) + +```python +from pythainlp.sentiment import sentiment +sentiment(str) +``` + +รับค่า str ส่งออกเป็น pos , neg + +### Util + +การใช้งาน + +```python +from pythainlp.util import * +``` + +#### ngrams + +สำหรับสร้าง n-grams + +```python +ngrams(token,num) +``` + +- token คือ list +- num คือ จำนวน ngrams + +#### bigrams + +สำหรับสร้าง bigrams + +```python +bigrams(token) +``` + +- token คือ list + +#### trigram + +สำหรับสร้าง trigram + +```python +trigram(token) +``` + +- token คือ list + +#### normalize + +ซ่อมข้อความภาษาไทย เช่น กี่่่ ไปเป็น กี่ + +```python +normalize(text) +``` + +**ตัวอย่าง** + +```python +>>> print(normalize("เเปลก")=="แปลก") # เ เ ป ล ก กับ แปลก +True +``` + +#### listtext_num2num + +แปลง list ข้อความตัวเลขในภาษาไทยให้เป็นตัวเลข + +```python +listtext_num2num(list) +``` + +**ตัวอย่าง** + +```python +>>> listtext_num2num(['หก','ล้าน','หกแสน','หกหมื่น','หกพัน','หกร้อย','หกสิบ','หก']) +6666666 +``` + +### Corpus + +#### WordNet ภาษาไทย + +เรียกใช้งาน + +```python +from pythainlp.corpus import wordnet +``` + +**การใช้งาน** + +API เหมือนกับ NLTK โดยรองรับ API ดังนี้ + +- wordnet.synsets(word) +- wordnet.synset(name_synsets) +- wordnet.all_lemma_names(pos=None, lang="tha") +- wordnet.all_synsets(pos=None) +- wordnet.langs() +- wordnet.lemmas(word,pos=None,lang="tha") +- wordnet.lemma(name_synsets) +- wordnet.lemma_from_key(key) +- wordnet.path_similarity(synsets1,synsets2) +- wordnet.lch_similarity(synsets1,synsets2) +- wordnet.wup_similarity(synsets1,synsets2) +- wordnet.morphy(form, pos=None) +- wordnet.custom_lemmas(tab_file, lang) + +**ตัวอย่าง** + +```python +>>> from pythainlp.corpus import wordnet +>>> print(wordnet.synsets('หนึ่ง')) +[Synset('one.s.05'), Synset('one.s.04'), Synset('one.s.01'), Synset('one.n.01')] +>>> print(wordnet.synsets('หนึ่ง')[0].lemma_names('tha')) +[] +>>> print(wordnet.synset('one.s.05')) +Synset('one.s.05') +>>> print(wordnet.synset('spy.n.01').lemmas()) +[Lemma('spy.n.01.spy'), Lemma('spy.n.01.undercover_agent')] +>>> print(wordnet.synset('spy.n.01').lemma_names('tha')) +['สปาย', 'สายลับ'] +``` + +#### stopword ภาษาไทย + +```python +from pythainlp.corpus import stopwords +stopwords = stopwords.words('thai') +``` + +#### ชื่อประเทศ ภาษาไทย + +```python +from pythainlp.corpus import country +country.get_data() +``` + +#### ตัววรรณยุกต์ในภาษาไทย + +```python +from pythainlp.corpus import tone +tone.get_data() +``` + +#### ตัวพยัญชนะในภาษาไทย + +```python +from pythainlp.corpus import alphabet +alphabet.get_data() +``` + +#### รายการคำในภาษาไทย + +```python +from pythainlp.corpus.thaiword import get_data # ข้อมูลเก่า +get_data() +from pythainlp.corpus.newthaiword import get_data # ข้อมูลใหม่ +get_data() +``` + +#### provinces + +สำหรับจัดการชื่อจังหวัดในประเทศไทย + +##### get_data + +รับข้อมูลชื่อจังหวัดในประเทศไทบ + +```python +get_data() +``` + +คือค่าออกมาเป็น list + +##### parsed_docs + +สำหรับใช้ Tag ชื่อจังหวัดในประเทศไทย + +```python +parsed_docs(text_list) +``` + +text_list คือ ข้อความภาษาไทยที่อยู่ใน list โดยผ่านการตัดคำมาแล้ว + +**ตัวอย่าง** + +```python +>>> d=['หนองคาย', 'เป็น', 'เมือง', 'น่าอยู่', 'อันดับ', 'ต้น', 'ๆ', 'ของ', 'โลก', 'นอกจากนี้', 'ยัง', 'มี', 'เชียงใหม่'] +>>> parsed_docs(d) +["[LOC : 'หนองคาย']", 'เป็น', 'เมือง', 'น่าอยู่', 'อันดับ', 'ต้น', 'ๆ', 'ของ', 'โลก', 'นอกจากนี้', 'ยัง', 'มี', "[LOC : 'เชียงใหม่']"] +``` + +#### ConceptNet + +เครื่องมือสำหรับ ConceptNet + +**ค้นหา edges** + +```python +edges(word,lang='th') +``` + +return dict + +#### TNC + +สำหรับใช้จัดการกับ Thai National Corpus (http://www.arts.chula.ac.th/~ling/TNC/index.php) + +##### word_frequency + +ใช้วัดความถี่ของคำ + +```python +word_frequency(word,domain='all') +``` + +word คือ คำ + +domain คือ หมวดหมู่ของคำ + +มีหมวดหมู่ดังนี้ + +- all +- imaginative +- natural-pure-science +- applied-science +- social-science +- world-affairs-history +- commerce-finance +- arts +- belief-thought +- leisure +- others + +เขียนโดย PyThaiNLP diff --git "a/docs/\340\270\241\340\270\265\340\270\255\340\270\260\340\271\204\340\270\243\340\271\203\340\270\253\340\270\241\340\271\210\340\271\203\340\270\231 PyThaiNLP 1.7.md" "b/docs/\340\270\241\340\270\265\340\270\255\340\270\260\340\271\204\340\270\243\340\271\203\340\270\253\340\270\241\340\271\210\340\271\203\340\270\231 PyThaiNLP 1.7.md" new file mode 100644 index 000000000..a6508a92d --- /dev/null +++ "b/docs/\340\270\241\340\270\265\340\270\255\340\270\260\340\271\204\340\270\243\340\271\203\340\270\253\340\270\241\340\271\210\340\271\203\340\270\231 PyThaiNLP 1.7.md" @@ -0,0 +1,14 @@ +# มีอะไรใหม่ใน PyThaiNLP 1.7 + +## สรุปประเด็นสำคัญ + +- ยกเลิกการสนับสนุน Python 2.7 อย่างเป็นทางการ +- เพิ่ม ULMFit utility เข้ามาใน PyThaiNLP +- ปรับปรุงระบบตัดคำใหม่ ทั้ง newmm และ mm +- thai2vec v0.2 +- sentiment analysis ตัวใหม่ทำงานด้วย Deep learning +- เพิ่ม thai2rom เป็น Thai Romanization ทำด้วย Deep learning ในระดับตัวอักษร + + + +กำลังปรับปรุง... \ No newline at end of file diff --git a/pythainlp/MetaSound.py b/pythainlp/MetaSound.py index c910910b0..1027462f0 100644 --- a/pythainlp/MetaSound.py +++ b/pythainlp/MetaSound.py @@ -11,13 +11,16 @@ import re def MetaSound(name): ''' - MetaSound(str) ภาษาไทย - - >>> MetaSound('รัก') - '501' - >>> MetaSound('ลัก') - '501' + Thai MetaSound + :param str name: thai text + :return: MetaSound for thai text + **Example**:: + >>> from pythainlp.MetaSound import MetaSound + >>> MetaSound('รัก') + '501' + >>> MetaSound('ลัก') + '501' ''' name1=list(name) count=len(name1) @@ -58,4 +61,4 @@ def MetaSound(name): return ''.join(name1) if __name__ == '__main__': print(MetaSound('รัก')) - print(MetaSound('ลัก')) \ No newline at end of file + print(MetaSound('ลัก')) diff --git a/pythainlp/__init__.py b/pythainlp/__init__.py index c50ee9d27..5475377b2 100644 --- a/pythainlp/__init__.py +++ b/pythainlp/__init__.py @@ -1,13 +1,8 @@ # -*- coding: utf-8 -*- from __future__ import absolute_import -__version__ = 1.6 -import six -if six.PY3: - """ - ไว้ใส่ความสามารถที่รองรับเฉพาะ Python 3.4+ เท่านั้น - """ - from pythainlp.sentiment import sentiment - from pythainlp.spell import spell +__version__ = 1.7 +from pythainlp.sentiment import sentiment +from pythainlp.spell import spell from pythainlp.romanization import romanization from pythainlp.tokenize import word_tokenize,sent_tokenize,tcc,etcc from pythainlp.rank import rank diff --git a/pythainlp/change.py b/pythainlp/change.py index 28aa6abd8..f9495369a 100644 --- a/pythainlp/change.py +++ b/pythainlp/change.py @@ -7,8 +7,10 @@ '~':u'%',u'`':u'_',u'@':u'๑',u'2':u'/',u'#':u'๒',u'3':u'-',u'$':u'๓',u'4':u'ภ',u'%':u'๔',u'5':u'ถ',u'^':u'ู',u'6':u'ุ',u'&':u'฿',u'7':u'ึ',u'*':u'๕',u'8':u'ค',u'(':u'๖',u'9':u'ต',u')':u'๗',u'0':u'จ',u'_':u'๘',u'-':u'ข',u'+':u'๙',u'=':u'ช'} # แก้ไขพิมพ์ภาษาไทยผิดภาษา def texttothai(data): - """เป็นคำสั่งแก้ไขข้อความที่พิมพ์ผิดภาษา ต้องการภาษาไทย แต่พิมพ์เป็นภาษาอังกฤษ - รับค่าเป็น ''str'' คืนค่าเป็น ''str''""" + """ + :param str data: Incorrect input language correction (Needs thai but input english) + :return: thai text + """ data = list(data) data2 = "" for a in data: @@ -21,8 +23,10 @@ def texttothai(data): return data2 # แก้ไขพิมพ์ภาษาอังกฤษผิดภาษา def texttoeng(data): - """เป็นคำสั่งแก้ไขข้อความที่พิมพ์ผิดภาษา ต้องการภาษาอังกฤษ แต่พิมพ์เป็นภาษาไทย - รับค่าเป็น ''str'' คืนค่าเป็น ''str''""" + """ + :param str data: Incorrect input language correction (Needs english but input thai) + :return: english text + """ data = list(data) data2 = "" dictdataeng= {v: k for k, v in six.iteritems(dictdata)} diff --git a/pythainlp/collation/__init__.py b/pythainlp/collation/__init__.py index c4833fef9..33687c763 100644 --- a/pythainlp/collation/__init__.py +++ b/pythainlp/collation/__init__.py @@ -5,20 +5,26 @@ try: import icu thkey = icu.Collator.createInstance(icu.Locale('th_TH')).getSortKey -except ImportError: +except ImportError: def thkey(word): cv = re.sub('[็-์]', '', word,re.U) # remove tone cv = re.sub('([เ-ไ])([ก-ฮ])', '\\2\\1', cv,re.U) # switch lead vowel tone = re.sub('[^็-์]', ' ', word,re.U) # just tone return cv+tone - -# เรียงลำดับข้อมูล list ภาษาไทย + def collation(data): - """เป็นคำสั่งเรียงลำดับข้อมูลใน ''list'' รับค่า list คืนค่าเป็น ''list''""" - return sorted(data, key=thkey) + """ + :param list data: a list of thai text + :return: a list of thai text, sorted alphabetically + **Example**:: + >>> from pythainlp.collation import * + >>> collation(['ไก่', 'เป็ด', 'หมู', 'วัว']) + ['ไก่', 'เป็ด', 'วัว', 'หมู'] + """ + return sorted(data, key=thkey) if __name__ == "__main__": a=collation(['ไก่','ไข่','ก','ฮา'])==['ก', 'ไก่', 'ไข่', 'ฮา'] print(a) print(collation(['หลาย','หญิง'])==['หญิง','หลาย']) - print(collation(['ไก่', 'เป็ด', 'หมู', 'วัว'])==['ไก่', 'เป็ด', 'วัว', 'หมู']) \ No newline at end of file + print(collation(['ไก่', 'เป็ด', 'หมู', 'วัว'])==['ไก่', 'เป็ด', 'วัว', 'หมู']) diff --git a/pythainlp/corpus/.ipynb_checkpoints/__init__-checkpoint.py b/pythainlp/corpus/.ipynb_checkpoints/__init__-checkpoint.py new file mode 100644 index 000000000..b8361f789 --- /dev/null +++ b/pythainlp/corpus/.ipynb_checkpoints/__init__-checkpoint.py @@ -0,0 +1,78 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,unicode_literals +from pythainlp.tools import get_path_db,get_path_data +from tinydb import TinyDB,Query +from future.moves.urllib.request import urlopen +from tqdm import tqdm +import requests +import os +import math +import requests +from nltk.corpus import names +#__all__ = ["thaipos", "thaiword","alphabet","tone","country","wordnet"] +path_db_=get_path_db() +def get_file(name): + db=TinyDB(path_db_) + temp = Query() + if len(db.search(temp.name==name))>0: + path= get_path_data(db.search(temp.name==name)[0]['file']) + db.close() + if not os.path.exists(path): + download(name) + return path +def download_(url, dst): + """ + @param: url to download file + @param: dst place to put the file + """ + file_size = int(urlopen(url).info().get('Content-Length', -1)) + if os.path.exists(dst): + first_byte = os.path.getsize(dst) + else: + first_byte = 0 + if first_byte >= file_size: + return file_size + header = {"Range": "bytes=%s-%s" % (first_byte, file_size)} + pbar = tqdm( + total=file_size, initial=first_byte, + unit='B', unit_scale=True, desc=url.split('/')[-1]) + req = requests.get(url, headers=header, stream=True) + with(open(get_path_data(dst), 'wb')) as f: + for chunk in req.iter_content(chunk_size=1024): + if chunk: + f.write(chunk) + pbar.update(1024) + pbar.close() + #return file_size +def download(name,force=False): + db=TinyDB(path_db_) + temp = Query() + data=requests.get("https://raw.githubusercontent.com/PyThaiNLP/pythainlp-corpus/master/db.json") + data_json=data.json() + if name in list(data_json.keys()): + temp_name=data_json[name] + print("Download : "+name) + if len(db.search(temp.name==name))==0: + print(name+" "+temp_name['version']) + download_(temp_name['download'],temp_name['file_name']) + db.insert({'name': name, 'version': temp_name['version'],'file':temp_name['file_name']}) + else: + if len(db.search(temp.name==name and temp.version==temp_name['version']))==0: + print("have update") + print("from "+name+" "+db.search(temp.name==name)[0]['version']+" update to "+name+" "+temp_name['version']) + yes_no="y" + if force==False: + yes_no=str(input("y or n : ")).lower() + if "y"==yes_no: + download_(temp_name['download'],temp_name['file_name']) + db.update({'version':temp_name['version']},temp.name==name) + else: + print("re-download") + print("from "+name+" "+db.search(temp.name==name)[0]['version']+" update to "+name+" "+temp_name['version']) + yes_no="y" + if force==False: + yes_no=str(input("y or n : ")).lower() + if "y"==yes_no: + download_(temp_name['download'],temp_name['file_name']) + db.update({'version':temp_name['version']},temp.name==name) + db.close() \ No newline at end of file diff --git a/pythainlp/corpus/__init__.py b/pythainlp/corpus/__init__.py index 18a3bfdb8..6dcc8678e 100644 --- a/pythainlp/corpus/__init__.py +++ b/pythainlp/corpus/__init__.py @@ -1,4 +1,86 @@ # -*- coding: utf-8 -*- from __future__ import absolute_import,unicode_literals -from nltk.corpus import names -#__all__ = ["thaipos", "thaiword","alphabet","tone","country","wordnet"] \ No newline at end of file +from pythainlp.tools import get_path_db,get_path_data +from tinydb import TinyDB,Query +from future.moves.urllib.request import urlopen +from tqdm import tqdm +import requests +import os +import requests +#__all__ = ["thaipos", "thaiword","alphabet","tone","country","wordnet"] +path_db_=get_path_db() +def get_file(name): + db=TinyDB(path_db_) + temp = Query() + if len(db.search(temp.name==name))>0: + path= get_path_data(db.search(temp.name==name)[0]['file']) + db.close() + if not os.path.exists(path): + download(name) + return path +def download_(url, dst): + """ + @param: url to download file + @param: dst place to put the file + """ + file_size = int(urlopen(url).info().get('Content-Length', -1)) + if os.path.exists(dst): + first_byte = os.path.getsize(dst) + else: + first_byte = 0 + if first_byte >= file_size: + return file_size + header = {"Range": "bytes=%s-%s" % (first_byte, file_size)} + pbar = tqdm( + total=file_size, initial=first_byte, + unit='B', unit_scale=True, desc=url.split('/')[-1]) + req = requests.get(url, headers=header, stream=True) + with(open(get_path_data(dst), 'wb')) as f: + for chunk in req.iter_content(chunk_size=1024): + if chunk: + f.write(chunk) + pbar.update(1024) + pbar.close() + #return file_size +def download(name,force=False): + db=TinyDB(path_db_) + temp = Query() + data=requests.get("https://raw.githubusercontent.com/PyThaiNLP/pythainlp-corpus/master/db.json") + data_json=data.json() + if name in list(data_json.keys()): + temp_name=data_json[name] + print("Download : "+name) + if len(db.search(temp.name==name))==0: + print(name+" "+temp_name['version']) + download_(temp_name['download'],temp_name['file_name']) + db.insert({'name': name, 'version': temp_name['version'],'file':temp_name['file_name']}) + else: + if len(db.search(temp.name==name and temp.version==temp_name['version']))==0: + print("have update") + print("from "+name+" "+db.search(temp.name==name)[0]['version']+" update to "+name+" "+temp_name['version']) + yes_no="y" + if force==False: + yes_no=str(input("y or n : ")).lower() + if "y"==yes_no: + download_(temp_name['download'],temp_name['file_name']) + db.update({'version':temp_name['version']},temp.name==name) + else: + print("re-download") + print("from "+name+" "+db.search(temp.name==name)[0]['version']+" update to "+name+" "+temp_name['version']) + yes_no="y" + if force==False: + yes_no=str(input("y or n : ")).lower() + if "y"==yes_no: + download_(temp_name['download'],temp_name['file_name']) + db.update({'version':temp_name['version']},temp.name==name) + db.close() +def remove(name): + db=TinyDB(path_db_) + temp = Query() + data=db.search(temp.name==name) + if len(data)>0: + path=get_file(name) + os.remove(path) + db.remove(temp.name==name) + return True + return False \ No newline at end of file diff --git a/pythainlp/corpus/new-thaidict.txt b/pythainlp/corpus/new-thaidict.txt index 785017cc0..920cfbb91 100755 --- a/pythainlp/corpus/new-thaidict.txt +++ b/pythainlp/corpus/new-thaidict.txt @@ -53100,8 +53100,6 @@ แค่นไค้ แคฝรั่ง แคลงคลาง -คำศัพท์หมวดอักษร "ค" -๑,๕๑๐ รายการ โคจร- โคจรคาม โคตร- diff --git a/pythainlp/corpus/pt_tagger_1.dill b/pythainlp/corpus/pt_tagger_1.dill new file mode 100644 index 000000000..bcfcf344a Binary files /dev/null and b/pythainlp/corpus/pt_tagger_1.dill differ diff --git a/pythainlp/corpus/stopwords-th1.txt b/pythainlp/corpus/stopwords-th1.txt deleted file mode 100644 index 5a8dc03f7..000000000 --- a/pythainlp/corpus/stopwords-th1.txt +++ /dev/null @@ -1,116 +0,0 @@ -กล่าว -กว่า -กัน -กับ -การ -ก็ -ก่อน -ขณะ -ขอ -ของ -ขึ้น -คง -ครั้ง -ความ -คือ -จะ -จัด -จาก -จึง -ช่วง -ซึ่ง -ดัง -ด้วย -ด้าน -ตั้ง -ตั้งแต่ -ตาม -ต่อ -ต่าง -ต่างๆ -ต้อง -ถึง -ถูก -ถ้า -ทั้ง -ทั้งนี้ -ทาง -ที่ -ที่สุด -ทุก -ทํา -ทําให้ -นอกจาก -นัก -นั้น -นี้ -น่า -นํา -บาง -ผล -ผ่าน -พบ -พร้อม -มา -มาก -มี -ยัง -รวม -ระหว่าง -รับ -ราย -ร่วม -ลง -วัน -ว่า -สุด -ส่ง -ส่วน -สําหรับ -หนึ่ง -หรือ -หลัง -หลังจาก -หลาย -หาก -อยาก -อยู่ -อย่าง -ออก -อะไร -อาจ -อีก -เขา -เข้า -เคย -เฉพาะ -เช่น -เดียว -เดียวกัน -เนื่องจาก -เปิด -เปิดเผย -เป็น -เป็นการ -เพราะ -เพื่อ -เมื่อ -เรา -เริ่ม -เลย -เห็น -เอง -แต่ -แบบ -แรก -และ -แล้ว -แห่ง -โดย -ใน -ให้ -ได้ -ไป -ไม่ -ไว้ -้ง \ No newline at end of file diff --git a/pythainlp/corpus/stopwords-th2.txt b/pythainlp/corpus/stopwords-th2.txt deleted file mode 100644 index 434ddacac..000000000 --- a/pythainlp/corpus/stopwords-th2.txt +++ /dev/null @@ -1,114 +0,0 @@ -ไว้ -ไม่ -ไป -ได้ -ให้ -ใน -โดย -แห่ง -แล้ว -และ -แรก -แบบ -แต่ -เอง -เห็น -เลย -เริ่ม -เรา -เมื่อ -เพื่อ -เพราะ -เป็นการ -เป็น -เปิดเผย -เปิด -เนื่องจาก -เดียวกัน -เดียว -เช่น -เฉพาะ -เคย -เข้า -เขา -อีก -อาจ -อะไร -ออก -อย่าง -อยู่ -อยาก -หาก -หลาย -หลังจาก -หลัง -หรือ -หนึ่ง -ส่วน -ส่ง -สุด -สําหรับ -ว่า -วัน -ลง -ร่วม -ราย -รับ -ระหว่าง -รวม -ยัง -มี -มาก -มา -พร้อม -พบ -ผ่าน -ผล -บาง -น่า -นี้ -นํา -นั้น -นัก -นอกจาก -ทุก -ที่สุด -ที่ -ทําให้ -ทํา -ทาง -ทั้งนี้ -ถ้า -ถูก -ถึง -ต้อง -ต่างๆ -ต่าง -ต่อ -ตาม -ตั้งแต่ -ตั้ง -ด้าน -ด้วย -ดัง -ซึ่ง -ช่วง -จึง -จาก -จัด -จะ -คือ -ความ -ครั้ง -คง -ขึ้น -ของ -ขอ -ขณะ -ก่อน -ก็ -การ -กับ -กัน -กว่า -กล่าว diff --git a/pythainlp/corpus/stopwords-th3.txt b/pythainlp/corpus/stopwords-th3.txt deleted file mode 100644 index e4cb167dc..000000000 --- a/pythainlp/corpus/stopwords-th3.txt +++ /dev/null @@ -1,322 +0,0 @@ -เกิน -เกินๆ -เกี่ยวกัน -เกี่ยวกับ -เกี่ยวของ -เกี่ยวเนื่อง -เกี่ยวๆ -เกือบ -เกือบจะ -เกือบๆ -แก -แก -แกไข -ใกล -ใกลๆ -ไกล -ไกลๆ -ขณะ -ขณะเดียวกัน -ขณะใด -ขณะใดๆ -ขณะที่ -ขณะนั้น -ขณะนี้ -ขณะหนึ่ง -ขวาง -ขวางๆ -ขอ -ของ -ขั้น -ใคร -ใคร -ใครจะ -ใครๆ -งาย -งายๆ -ไง -จง -จด -จน -จนกระทั่ง -จนกวา -จนขณะนี้ -จนตลอด -จนถึง -จนทั่ว -จนบัดนี้ -จนเมื่อ -จนแม -จนแมน -จรด -จรดกับ -จริง -จริงจัง -จริงๆ -จริงๆจังๆ -จวน -จวนจะ -จวนเจียน -จวบ -ซึ่งก็ -ซึ่งก็คือ -ซึ่งกัน -ซึ่งกันและกัน -ซึ่งไดแก -ซึ่งๆ -ณ -ดวย -ดวยกัน -ดวยเชนกัน -ดวยที่ -ดวยประการฉะนี้ -ดวยเพราะ -ดวยวา -ดวยเหตุที่ -ดวยเหตุนั้น -ดวยเหตุนี้ -ดวยเหตุเพราะ -ดวยเหตุวา -ดวยเหมือนกัน -ดัง -ดั่ง -ดังกลาว -ดังกับ -ดั่งกับ -ดังกับวา -ดั่งกับวา -ดังเกา -ดั่งเกา -ดังเคย -ใดๆ -ได -ไดแก -ไดแต -ไดที่ -ไดมา -ไดรับ -ตน -ตนเอง -ตนฯ -ตรง -ตรงๆ -ตลอด -ตลอดกาล -ตลอดกาลนาน -ตลอดจน -ตลอดถึง -ตลอดทั้ง -ตลอดทั่ว -ตลอดทั่วถึง -ตลอดทั่วทั้ง -ตลอดป -ตลอดไป -ตลอดมา -ตลอดระยะเวลา -ตลอดวัน -ตลอดเวลา -ตลอดศก -ตอ -ตอกัน -ถึง -ถึงแก -ถึงจะ -ถึงบัดนั้น -ถึงบัดนี้ -ถึงเมื่อ -ถึงเมื่อใด -ถึงเมื่อไร -ถึงแม -ถึงแมจะ -ถึงแมวา -ถึงอยางไร -ถือ -ถือวา -ถูก -ถูกตอง -ถูกๆ -เถอะ -เถิด -ทรง -ทวา -ทั้ง -ทั้งคน -ทั้งตัว -ทั้งที -ทั้งที่ -ทั้งนั้น -ทั้งนั้นดวย -ทั้งนั้นเพราะ -ทั้งนี้ -นอก -นอกจาก -นอกจากที่ -นอกจากนั้น -นอกจากนี้ -นอกจากวา -นอกนั้น -นอกเหนือ -นอกเหนือจาก -นอย -นอยกวา -นอยๆ -นะ -นะ -นัก -นักๆ -นั่น -นั้น -นั่นไง -นั่นเปน -นั่นแหละ -นั่นเอง -นั้นๆ -นับ -นับจากนั้น -นับจากนี้ -นับตั้งแต -นับแต -นับแตที่ -นับแตนั้น -เปนตน -เปนตนไป -เปนตนมา -เปนแต -เปนแตเพียง -เปนที -เปนที่ -เปนที่สุด -เปนเพราะ -เปนเพราะวา -เปนเพียง -เปนเพียงวา -เปนเพื่อ -เปนอัน -เปนอันมาก -เปนอันวา -เปนอันๆ -เปนอาทิ -เปนๆ -เปลี่ยน -เปลี่ยนแปลง -เปด -เปดเผย -ไป -ไป -ผาน -ผานๆ -ผิด -ผิดๆ -ผู -เพียงเพื่อ -เพียงไร -เพียงไหน -เพื่อ -เพื่อที่ -เพื่อที่จะ -เพื่อวา -เพื่อให -ภาค -ภาคฯ -ภาย -ภายใต -ภายนอก -ภายใน -ภายภาค -ภายภาคหนา -ภายหนา -ภายหลัง -มอง -มองวา -มัก -มักจะ -มัน -มันๆ -มั้ย -มั้ยนะ -มั้ยนั่น -มั้ยเนี่ย -มั้ยละ -มา -ยืนนาน -ยืนยง -ยืนยัน -ยืนยาว -เยอะ -เยอะแยะ -เยอะๆ -แยะ -แยะๆ -รวด -รวดเร็ว -รวม -รวม -รวมกัน -รวมกัน -รวมดวย -รวมดวย -รวมถึง -รวมทั้ง -รวมมือ -รวมๆ -ระยะ -ระยะๆ -ระหวาง -รับ -รับรอง -รึ -รึวา -รือ -รือวา -สิ้นกาลนาน -สืบเนื่อง -สุด -สุดๆ -สู -สูง -สูงกวา -สูงสง -สูงสุด -สูงๆ -เสมือนกับ -เสมือนวา -เสร็จ -เสร็จกัน -เสร็จแลว -เสร็จสมบูรณ -เสร็จสิ้น -เสีย -เสียกอน -เสียจน -เสียจนกระทั่ง -เสียจนถึง -เสียดวย -เสียนั่น -เสียนั่นเอง -เสียนี่ -เสียนี่กระไร -เสียยิ่ง -เสียยิ่งนัก -เสียแลว -ใหญๆ -ใหดี -ใหแด -ใหไป -ใหม -ใหมา -ใหมๆ -ไหน -ไหนๆ -อดีต -อนึ่ง -อยาก -อยาง -อยางเชน -อยางดี -อยางเดียว -อยางใด -อยางที่ -อยางนอย -อยางนั้น -อยางนี้ -อยางโนน diff --git a/pythainlp/corpus/stopwords-th4.txt b/pythainlp/corpus/stopwords-th4.txt deleted file mode 100644 index 2424dac00..000000000 --- a/pythainlp/corpus/stopwords-th4.txt +++ /dev/null @@ -1,887 +0,0 @@ -ก็ -ก็คือ -ก็แค่ -ก็จะ -ก็ดี -ก็ได้ -ก็ต่อเมื่อ -ก็ตาม -ก็ตามแต่ -ก็ตามที -ก็แล้วแต่ -กระทั่ง -กระทำ -กระนั้น -กระผม -กลับ -กล่าว -กล่าวคือ -กลุ่ม -กลุ่มก้อน -กลุ่มๆ -กว่า -กว้าง -กว้างขวาง -กว้างๆ -ก่อน -ก่อนหน้า -ก่อนหน้านี้ -ก่อนๆ -กัน -กันดีกว่า -กันดีไหม -กันเถอะ -กันนะ -กันและกัน -กันไหม -กันเอง -กับ -การ -กำลัง -กำลังจะ -กำหนด -กู -เก็บ -เกิด -เกิน -เกินๆ -เกี่ยวกัน -เกี่ยวกับ -เกี่ยวข้อง -เกี่ยวเนื่อง -เกี่ยวๆ -เกือบ -เกือบจะ -เกือบๆ -แก -แก่ -แก้ไข -ใกล้ -ใกล้ๆ -ไกล -ไกลๆ -ขณะ -ขณะเดียวกัน -ขณะใด -ขณะใดๆ -ขณะที่ -ขณะนั้น -ขณะนี้ -ขณะหนึ่ง -ขวาง -ขวางๆ -ขอ -ของ -ขั้น -ข้า -ข้าง -ข้างเคียง -ข้างต้น -ข้างบน -ข้างล่าง -ข้างๆ -ขาด -ข้าพเจ้า -ข้าฯ -ขึ้น -เขา -เข้า -เข้าใจ -เขียน -คง -คงจะ -คงอยู่ -ครบ -ครบครัน -ครบถ้วน -ครั้ง -ครั้งกระนั้น -ครั้งก่อน -ครั้งครา -ครั้งคราว -ครั้งใด -ครั้งที่ -ครั้งนั้น -ครั้งนี้ -ครั้งละ -ครั้งหนึ่ง -ครั้งหลัง -ครั้งหลังสุด -ครั้งไหน -ครั้งๆ -ครัน -ครับ -ครา -คราใด -คราที่ -ครานั้น -ครานี้ -คราหนึ่ง -คราไหน -คราว -คราวก่อน -คราวใด -คราวที่ -คราวนั้น -คราวนี้ -คราวโน้น -คราวละ -คราวหน้า -คราวหนึ่ง -คราวหลัง -คราวไหน -คราวๆ -คล้าย -คล้ายกัน -คล้ายกันกับ -คล้ายกับ -คล้ายกับว่า -คล้ายว่า -ควร -ความ -ค่อน -ค่อนข้าง -ค่อนข้างจะ -ค่อยไปทาง -ค่อนมาทาง -ค่อย -ค่อยๆ -คะ -ค่ะ -คำ -คิด -คิดว่า -คือ -คุณ -คุณๆ -เคย -เคยๆ -แค่ -แค่จะ -แค่นั้น -แค่นี้ -แค่เพียง -แค่ว่า -แค่ไหน -ใคร -ใคร่ -ใคร่จะ -ง่าย -ง่ายๆ -ไง -จง -จด -จนกระทั่ง -จนกว่า -จนขณะนี้ -จนตลอด -จนถึง -จนทั่ว -จนบัดนี้ -จนเมื่อ -จนแม้ -จนแม้น -จรด -จรดกับ -จริง -จริงจัง -จริงๆ -จริงๆ -จังๆ -จวน -จวนจะ -จวนเจียน -จวบ -จวบกับ -จวบจน -จะ -จ้ะ -จ๊ะ -จะได้ -จัง -จังๆ -จัด -จัดการ -จัดงาน -จัดแจง -จัดตั้ง -จัดทำ -จัดหา -จัดให้ -จับ -จ้า -จ๋า -จาก -จากนั้น -จากนี้ -จากนี้ไป -จำ -จำเป็น -จำพวก -จึง -จึงจะ -จึงเป็น -จู่ๆ -ฉะนั้น -ฉะนี้ -ฉัน -เฉกเช่น -เฉพาะ -เฉย -เฉยๆ -ไฉน -ช่วง -ช่วงก่อน -ช่วงต่อไป -ช่วงถัดไป -ช่วงท้าย -ช่วงที่ -ช่วงนั้น -ช่วงนี้ -ช่วงระหว่าง -ช่วงแรก -ช่วงหน้า -ช่วงหลัง -ช่วงๆ -ช่วย -ช้า -ช้านาน -ชาว -ช้าๆ -เช่น -เช่นก่อน -เช่นกัน -เช่นเคย -เช่นดัง -เช่นดังก่อน -เช่นดังเก่า -เช่นดังที่ -เช่นดังว่า -เช่นเดียวกัน -เช่นเดียวกับ -เช่นใด -เช่นที่ -เช่นที่เคย -เช่นที่ว่า -เช่นนั้น -เช่นนั้นเอง -เช่นนี้ -เช่นเมื่อ -เช่นไร -เชื่อ -เชื่อถือ -เชื่อมั่น -เชื่อว่า -ใช่ -ใช่ไหม -ใช้ -ซะ -ซะก่อน -ซะจน -ซะจนกระทั่ง -ซะจนถึง -ซึ่ง -ซึ่งก็ -ซึ่งก็คือ -ซึ่งกัน -ซึ่งกันและกัน -ซึ่งได้แก่ -ซึ่งๆ -ณ -ด้วย -ด้วยกัน -ด้วยเช่นกัน -ด้วยที่ -ด้วยเพราะ -ด้วยว่า -ด้วยเหตุที่ -ด้วยเหตุนั้น -ด้วยเหตุนี้ -ด้วยเหตุเพราะ -ด้วยเหตุว่า -ด้วยเหมือนกัน -ดัง -ดั่ง -ดังกล่าว -ดังกับ -ดั่งกับ -ดังกับว่า -ดั่งกับว่า -ดังเก่า -ดั่งเก่า -ดังเคย -ดั่งเคย -ต่าง -ต่างก็ -ต่างหาก -ต่างๆ -ตาม -ตามด้วย -ตามแต่ -ตามที่ -ตามๆ -เต็มไปด้วย -เต็มไปหมด -เต็มๆ -แต่ -แต่ก็ -แต่ก่อน -แต่จะ -แต่เดิม -แต่ต้อง -แต่ถ้า -แต่ทว่า -แต่ที่ -แต่นั้น -แต่เพียง -แต่เมื่อ -แต่ไร -แต่ละ -แต่ว่า -แต่ไหน -แต่อย่างใด -โต -โตๆ -ใต้ -ถ้า -ถ้าจะ -ถ้าหาก -ถึง -ถึงแก่ -ถึงจะ -ถึงบัดนั้น -ถึงบัดนี้ -ถึงเมื่อ -ถึงเมื่อใด -ถึงเมื่อไร -ถึงแม้ -ถึงแม้จะ -ถึงแม้ว่า -ถึงอย่างไร -ถือ -ถือว่า -ถูก -ถูกต้อง -ถูกๆ -เถอะ -เถิด -ทรง -ทว่า -ทั้ง -ทั้งคน -ทั้งตัว -ทั้งที่ -ทั้งนั้น -ทั้งนั้นด้วย -ทั้งนั้นเพราะ -ทั้งนี้ -ทั้งปวง -ทั้งเป็น -ทั้งมวล -ทั้งสิ้น -ทั้งหมด -ทั้งหลาย -ทั้งๆ -ทั้งๆ -ที่ -ทัน -ทันใดนั้น -ทันที -ทันทีทันใด -ทั่ว -ท -าไม -ท -าไร -ท -าให้ -ท -าๆ -ที -ที่ -ที่จริง -ที่ซึ่ง -ทีเดียว -ทีใด -ที่ใด -ที่ได้ -ทีเถอะ -ที่แท้ -ที่แท้จริง -ที่นั้น -ที่นี้ -ทีไร -ทีละ -ที่ละ -ที่แล้ว -ที่ว่า -ที่สุด -ที่แห่งนั้น -ที่ไหน -ทีๆ -ที่ๆ -ทุก -ทุกคน -ทุกครั้ง -ทุกครา -ทุกคราว -ทุกชิ้น -ทุกตัว -ทุกทาง -ทุกที -ทุกที่ -ทุกเมื่อ -ทุกวัน -ทุกวันนี้ -ทุกสิ่ง -ทุกหน -ทุกแห่ง -ทุกอย่าง -ทุกอัน -ทุกๆ -เท่า -เท่ากัน -เท่ากับ -เท่าใด -เท่าที่ -เท่านั้น -เท่านี้ -เท่าไร -เท่าไหร่ -แท้ -แท้จริง -เธอ -นอก -นอกจาก -นอกจากที่ -นอกจากนั้น -นอกจากนี้ -นอกจากว่า -นอกนั้น -นอกเหนือ -น้อย -น้อยกว่า -น้อยๆ -นะ -น่ะ -นัก -นั่น -นั้นไว -นับจากนั้น -นับจากนี้ -นับแต่นี้ -น่า -นาง -นางสาว -น่าจะ -นาน -นานๆ -นาย -นำ -นำพา -นำมา -นิด -นิดหน่อย -นิดๆ -นี่ -นี้ -นี่ไง -นี่นา -นี่แน่ะ -นี่แหละ -นี้แหล่ -นี่เอง -นี้เอง -นู่น -นู้น -เน้น -เนี่ย -เนี่ยเอง -ใน -ในช่วง -ในที่ -ในเมื่อ -ในระหว่าง -บน -บอก -บอกแล้ว -บอกว่า -บ่อย -บ่อยกว่า -บ่อยครั้ง -บ่อยๆ -บัดดล -บัดเดี๋ยวนี้ -บัดนั้น -บัดนี้ -บาง -บ้าง -บางกว่า -บางขณะ -บางครั้ง -บางครา -บางคราว -บางที -บางที่ -บางแห่ง -บางๆ -แบบ -ปฏิบัติ -ประกอบ -ประการ -ประการฉะนี้ -ประการใด -ประการหนึ่ง -ประมาณ -ประสบ -ปรับ -ปรากฏ -ปรากฏว่า -ปัจจุบัน -ปิด -เป็น -เป็นด้วย -เป็นดัง -เป็นต้น -เป็นแต่ -เป็นเพื่อ -เป็นอัน -เป็นอันมาก -เป็นอาทิ -เปลี่ยน -เปลี่ยนแปลง -ไป -ผ่าน -ผ่านๆ -ผิด -ผิดๆ -ผู้ -ผู้ใด -เผื่อ -เผื่อจะ -เผื่อที่ -เผื่อว่า -ฝ่าย -ฝ่ายใด -พบ -พบว่า -พยายาม -พร้อม -พร้อมกัน -พร้อมกับ -พร้อมด้วย -พร้อมทั้ง -พร้อมที่ -พร้อมเพียง -พวก -พวกกัน -พวกกู -พวกแก -พวกเขา -พวกคุณ -พวกฉัน -พวกท่าน -พวกที่ -พวกเธอ -พวกนั้น -พวกนี้ -พวกนู้น -พวกโน้น -พวกมัน -พวกมึง -พอ -พอกัน -พอควร -พอจะ -พอดี -พอตัว -พอที -พอที่ -พอเพียง -พอแล้ว -พอสม -พอสมควร -พอเหมาะ -พอๆ -พา -พึง -พึ่ง -พื้นๆ -พูด -เพราะ -เพราะฉะนั้น -เพราะว่า -เพิ่ง -เพิ่งจะ -เพิ่ม -เพิ่มเติม -เพียง -เพียงแค่ -เพียงใด -เพียงแต่ -เพียงพอ -เพียงเพราะ -เพียงเพื่อ -เพียงไหน -เพื่อ -เพื่อที่ -เพื่อว่า -เพื่อให้ -ภาค -ภาย -ภายใต้ -ภายนอก -ภายหลัง -มอง -มองว่า -มัก -มักจะ -มัน -มั๊ย -มา -มาก -มากกว่า -มากมาย -มิ -มิฉะนั้น -มิใช่ -มิได้ -มี -มีแต่ -มึง -มุ่ง -มุ่งเน้น -มุ่งหมาย -เมื่อ -เมื่อก่อน -เมื่อครั้ง -เมื่อครั้งก่อน -เมื่อคราวก่อน -เมื่อคราวที่ -เมื่อคราว -เมื่อคืน -เมื่อเช้า -เมื่อใด -เมื่อนั้น -เมื่อนี้ -เมื่อเย็น -เมื่อไร -เมื่อวันวาน -เมื่อวาน -เมื่อไหร่ -แม้ -แม้กระทั่ง -แม้แต่ -แม้นว่า -แม้ว่า -ไม่ -ไม่ค่อย -ไม่ค่อยจะ -ไม่ค่อยเป็น -ไม่ใช่ -ไม่เป็นไร -ไม่ว่า -ยก -ยกให้ -ยอม -ยอมรับ -ย่อม -ย่อย -ยัง -ยังคง -ยังงั้น -ยังงี้ -ยังโง้น -ยังไง -ยังจะ -ยังแต่ -ยาก -ยาว -ยาวนาน -ยิ่ง -ยิ่งกว่า -ยิ่งขึ้น -ยิ่งขึ้นไป -ยิ่งจน -ยิ่งจะ -ยิ่งนัก -ยิ่งเมื่อ -ยิ่งแล้ว -ยิ่งใหญ่ -เยอะ -เยอะแยะ -แยะ -รวด -รวดเร็ว -รวม -ร่วม -รวมกัน -ร่วมกัน -รวมด้วย -ร่วมด้วย -รวมถึง -รวมทั้ง -ระยะ -ระหว่าง -รับ -รึ -รือ -รือว่า -เร็ว -เร็วๆ -เรา -เราๆ -เริ่ม -เรียก -เรียบ -เรื่อย -เรื่อยๆ -ไร -ล้วน -ล้วนจน -ล้วนแต่ -ละ -ล่าสุด -เล็ก -เล็กน้อย -เล็กๆ -เลย -เล่าว่า -แล้ว -แล้วกัน -แล้วแต่ -แล้วเสร็จ -วันใด -วันนั้น -วันนี้ -วันไหน -สบาย -สมัย -สมัยก่อน -สมัยนั้น -สมัยนี้ -สมัยโน้น -ส่วน -ส่วนเกิน -ส่วนด้อย -ส่วนดี -ส่วนใด -ส่วนที่ -ส่วนน้อย -ส่วนนั้น -ส่วนมาก -ส่วนใหญ่ -สั้น -สั้นๆ -สามารถ -สำคัญ -สิ่ง -สิ่งใด -สิ่งนั้น -สิ่งนี้ -สิ่งไหน -สิ้น -สุด -เสร็จ -เสร็จแล้ว -เสียจน -เสียด้วย -เสียนี่ -เสียแล้ว -แสดง -แสดงว่า -หน -หนอ -หนอย -หน่อย -หมด -หมดกัน -หมดสิ้น -หรือไง -หรือเปล่า -หรือไม่ -หรือยัง -หรือไร -หลังจาก -หาก -หากแม้ -หากแม้น -หากแม้นว่า -หากว่า -หาความ -หาใช่ -หารือ -เหตุ -เหตุผล -เหตุนั้น -เหตุนี้ -เหตุไร -เห็นแก่ -เห็นควร -เห็นจะ -เห็นว่า -เหลือ -เหลือเกิน -เหล่า -เหล่านั้น -เหล่านี้ -แห่ง -แห่งใด -แห่งนั้น -แห่งนี้ -แห่งโน้น -แห่งไหน -แหละ -ให้ -ให้แก่ -ใหญ่ -ใหญ่โต -ไหน -ไหนๆ -อดีต -อนึ่ง -อยาก -อย่าง -อย่างเช่น -อย่างดี -อย่างเดียว -อย่างใด -อย่างที่ -อย่างน้อย -อย่างนั้น -อย่างนี้ -อย่างโน้น -อย่างมาก -อย่างยิ่ง -อย่างไร -อย่างไรก็ -อย่างไรก็ได้ -อย่างไรเสีย -อย่างละ -อย่างหนึ่ง -อย่างไหน -อย่างๆ -อัน -อันจะ -อันใด -อันได้แก่ -อันที่ -อันที่จริง -อันที่จะ -อันเนื่องมาจาก -อันละ -อันไหน -อันๆ -อาจ -อาจจะ -อาจเป็น -อาจเป็นด้วย -อีก -อื่น -อื่นๆ -เอง -เอ็ง -เอา -ฯ -ฯล -ฯลฯ \ No newline at end of file diff --git a/pythainlp/corpus/thaiword.txt b/pythainlp/corpus/thaiword.txt index 8ecdd335c..985554f0f 100644 --- a/pythainlp/corpus/thaiword.txt +++ b/pythainlp/corpus/thaiword.txt @@ -7582,7 +7582,6 @@ คอวี คอส. คอสติกโซดา -คอสอง คอสะพาน คอสูง คอหนัง @@ -9873,7 +9872,6 @@ จ๋า จ๋ำหนับ จําเลย -ฉ ฉก ฉกจวัก ฉกฉวย @@ -9884,7 +9882,6 @@ ฉกาจ ฉกาจฉกรรจ์ ฉกามาพจร -ฉง ฉงชิ่ง ฉงน ฉงนสนเท่ห์ @@ -10077,7 +10074,6 @@ ฉ้อฉล ฉ้อราษฎร์บังหลวง ฉ้อโกง -ช ช.ค. ช.ค.บ. ช.บ.ร. @@ -11755,7 +11751,6 @@ ฑากะ ฑาหก ฑาหะ -ณ ณรงค์ ณัฏฐพัชร วิพัธครตระกูล ณัฏฐ์ เทพหัสดิน @@ -14910,7 +14905,6 @@ ถ้ำชา ถ้ำมอง ถ้ำยาดม -ท ท.ญ. ท.ทบ. ท.บ. @@ -16553,7 +16547,6 @@ ทําปฏิกิริยา ทําเนียบ ทําไม -ธ ธ. ธ.ก.ส ธ.ก.ส. @@ -16799,7 +16792,6 @@ ธเรษตรี ธโนปจัย ธไนศวรรย์ -น น. น.ช. น.ญ. @@ -18323,7 +18315,6 @@ น้ำไว น้ำไหลไฟดับ นํ้า -บ บ. บ. เดอร์มอล คอนแซปท์ จำกัด บ.ก. @@ -22735,7 +22726,6 @@ ฝนชะลาน ฝนดาวตก ฝนตก -ฝนปรอย ฝนฟ้า ฝนลงเม็ด ฝนหลวง @@ -25881,7 +25871,6 @@ มหุดิฤกษ์ มหุรดี มหู -มอ มอก. มอคราม มอง @@ -29500,7 +29489,6 @@ ร้าวฉาน ร้าวราน ร๊ายร้าย -ฤ ฤกษณะ ฤกษ์ ฤกษ์ดี @@ -32301,7 +32289,6 @@ ษัษฏี ษัษฐะ ษัษฐี -ส ส.ก. ส.ค. ส.ค.1 @@ -32872,6 +32859,7 @@ สมผุส สมพง สมพงศ์ +สมพร สมพรปาก สมพล สมพัตสร @@ -37277,7 +37265,6 @@ ห้ำ ห้ำหั่น ห๊วยห่วย -อ อ. อ.จ. อ.ฐากูร เกิดแก้ว @@ -39353,8 +39340,6 @@ อิหลักอิเหลื่อ อิออน อิอิ -อิอิอิ -อิอิอิอิ อิเคลติคซิสม์ อิเลคโตรไลต์ อิเลคโทรนิคส์ @@ -51694,4 +51679,5 @@ ไฮไฟ ไฮไฟว์ ไฮไลต์ -ไฮ้ \ No newline at end of file +ไฮ้ +กรกนก diff --git a/pythainlp/corpus/ud_thai-pud_pt_tagger.dill b/pythainlp/corpus/ud_thai-pud_pt_tagger.dill new file mode 100644 index 000000000..d3fcf058a Binary files /dev/null and b/pythainlp/corpus/ud_thai-pud_pt_tagger.dill differ diff --git a/pythainlp/corpus/ud_thai-pud_unigram_tagger.dill b/pythainlp/corpus/ud_thai-pud_unigram_tagger.dill new file mode 100644 index 000000000..511f3100e Binary files /dev/null and b/pythainlp/corpus/ud_thai-pud_unigram_tagger.dill differ diff --git a/pythainlp/date/__init__.py b/pythainlp/date/__init__.py index 3fb125cd1..66dfe39ec 100644 --- a/pythainlp/date/__init__.py +++ b/pythainlp/date/__init__.py @@ -5,6 +5,9 @@ tz = pytz.timezone('Asia/Bangkok') def now(): + """ + :return: the current date with Thai month and Thai year. The month is spelled out in text, and the year is converted from AD to Thai years. (ie: 30 ตุลาคม 2560 20:45:30) + """ now1 = datetime.datetime.now(tz) month_name = 'x มกราคม กุมภาพันธ์ มีนาคม เมษายน พฤษภาคม มิถุนายน กรกฎาคม สิงหาคม กันยายน ตุลาคม พฤศจิกายน ธันวาคม'.split()[now1.month] thai_year = now1.year + 543 @@ -21,7 +24,7 @@ def reign_year_to_ad(reign_year,reign): ''' ปีรัชกาล แปลงเป็น ค.ศ. reign_year_to_ad(reign_year,reign) - reign_year - ปีที่ + reign_year - ปีที่ reign - รัชกาล ''' if int(reign)==10: @@ -35,4 +38,4 @@ def reign_year_to_ad(reign_year,reign): return ad # BE คือ พ.ศ. # AD คือ ค.ศ. -# AH ปีฮิจเราะห์ศักราชเป็นปีพุทธศักราช จะต้องบวกด้วย 1122 \ No newline at end of file +# AH ปีฮิจเราะห์ศักราชเป็นปีพุทธศักราช จะต้องบวกด้วย 1122 diff --git a/pythainlp/keywords/__init__.py b/pythainlp/keywords/__init__.py index 9b6f3c9a7..826593141 100644 --- a/pythainlp/keywords/__init__.py +++ b/pythainlp/keywords/__init__.py @@ -4,13 +4,10 @@ from pythainlp.rank import rank def find_keyword(word_list,lentext=3): ''' - ระบบค้นหาคำสำคัญ - หลักการ ลบ stopword ออกแล้ว นับจำนวนคำที่ส่งค่าออกมา - - find_keyword(word_list,lentext=3) - word_list คือ คำที่อยู่ใน list - lentext คือ จำนวนคำที่มีอยู่ใน list สำหรับใช้กำหนดค่าหา keyword ค่าเริ่มต้นคือ 3 + :param list word_list: a list of thai text + :param int lentext: a number of keyword + :return: dict ''' filtered_words = [word for word in word_list if word not in set(stopwords.words('thai'))] word_list=rank(filtered_words) - return {k:v for k, v in word_list.items() if v>=lentext} \ No newline at end of file + return {k:v for k, v in word_list.items() if v>=lentext} diff --git a/pythainlp/ner/__init__.py b/pythainlp/ner/__init__.py new file mode 100644 index 000000000..e9296bbea --- /dev/null +++ b/pythainlp/ner/__init__.py @@ -0,0 +1,125 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,unicode_literals +# NLP +import re +from pythainlp.tokenize import word_tokenize +from pythainlp.tag import pos_tag +from pythainlp.corpus import stopwords +thaicut="newmm" # ตัวตัดคำ +# CRF +try: + import sklearn_crfsuite +except ImportError: + from pythainlp.tools import install_package + install_package('sklearn-crfsuite') + import sklearn_crfsuite +# FILE +import glob +import codecs +from pythainlp.corpus import get_file,download + +stopwords = stopwords.words('thai') + + +def isThai(chr): # เช็คว่าเป็น char ภาษาไทย + cVal = ord(chr) + if(cVal >= 3584 and cVal <= 3711): + return True + return False +def isThaiWord(word): # เช็คว่าเป็นคำภาษาไทย + t=True + for i in word: + l=isThai(i) + if l!=True and i!='.': + t=False + break + return t + +def is_stopword(word): # เช็คว่าเป็นคำฟุ่งเฟือง + return word in stopwords +def doc2features(doc, i): + word = doc[i][0] + postag = doc[i][1] + # Features from current word + features={ + 'word.word': word, + 'word.stopword': is_stopword(word), + 'word.isthai':isThaiWord(word), + 'word.isspace':word.isspace(), + 'postag':postag, + 'word.isdigit()': word.isdigit() + } + if word.isdigit() and len(word)==5: + features['word.islen5']=True + if i > 0: + prevword = doc[i-1][0] + postag1 = doc[i-1][1] + features['word.prevword'] = prevword + features['word.previsspace']=prevword.isspace() + features['word.previsthai']=isThaiWord(prevword) + features['word.prevstopword']=is_stopword(prevword) + features['word.prepostag'] = postag1 + features['word.prevwordisdigit'] = prevword.isdigit() + else: + features['BOS'] = True # Special "Beginning of Sequence" tag + # Features from next word + if i < len(doc)-1: + nextword = doc[i+1][0] + postag1 = doc[i+1][1] + features['word.nextword'] = nextword + features['word.nextisspace']=nextword.isspace() + features['word.nextpostag'] = postag1 + features['word.nextisthai']=isThaiWord(nextword) + features['word.nextstopword']=is_stopword(nextword) + features['word.nextwordisdigit'] = nextword.isdigit() + else: + features['EOS'] = True # Special "End of Sequence" tag + return features + +class thainer: + def __init__(self): + """ + Thai NER + """ + self.data_path = get_file('thainer') + if self.data_path==None: + download('thainer') + self.data_path = get_file('thainer') + self.crf=sklearn_crfsuite.CRF( + algorithm='lbfgs', + c1=0.1, + c2=0.1, + max_iterations=500, + all_possible_transitions=True, + model_filename=self.data_path) + def get_ner(self,text,postag=True): + """ + Get NER from Thai NER. + + :param string text: thai text + :param boolean postag: get postag (True) or get not postag (False) + + :return: list NER. + + **Example**:: + >>> from pythainlp.ner import thainer + >>> ner=thainer() + >>> ner.get_ner("วันที่ 15 ก.ย. 61 ทดสอบระบบเวลา 14:49 น.") + [('วันที่', 'JSBR', 'O'), (' ', 'NCMN', 'O'), ('15', 'NCNM', 'B-DATE'), (' ', 'NCMN', 'I-DATE'), ('ก.ย.', 'CMTR', 'I-DATE'), (' ', 'NCMN', 'I-DATE'), ('61', 'NCNM', 'I-DATE'), (' ', 'NCMN', 'O'), ('ทดสอบ', 'VACT', 'O'), ('ระบบ', 'NCMN', 'O'), ('เวลา', 'NCMN', 'O'), (' ', 'NCMN', 'O'), ('14', 'NCNM', 'B-TIME'), (':', 'PUNC', 'I-TIME'), ('49', 'NCNM', 'I-TIME'), (' ', 'NCMN', 'I-TIME'), ('น.', 'CMTR', 'I-TIME')] + >>> ner.get_ner("วันที่ 15 ก.ย. 61 ทดสอบระบบเวลา 14:49 น.",postag=False) + [('วันที่', 'O'), (' ', 'O'), ('15', 'B-DATE'), (' ', 'I-DATE'), ('ก.ย.', 'I-DATE'), (' ', 'I-DATE'), ('61', 'I-DATE'), (' ', 'O'), ('ทดสอบ', 'O'), ('ระบบ', 'O'), ('เวลา', 'O'), (' ', 'O'), ('14', 'B-TIME'), (':', 'I-TIME'), ('49', 'I-TIME'), (' ', 'I-TIME'), ('น.', 'I-TIME')] + """ + self.word_cut=word_tokenize(text,engine=thaicut) + self.list_word=pos_tag(self.word_cut,engine='perceptron') + self.X_test = self.extract_features([(data,self.list_word[i][1]) for i,data in enumerate(self.word_cut)]) + self.y_=self.crf.predict_single(self.X_test) + if postag: + return [(self.word_cut[i],self.list_word[i][1],data) for i,data in enumerate(self.y_)] + else: + return [(self.word_cut[i],data) for i,data in enumerate(self.y_)] + def extract_features(self,doc): + return [doc2features(doc, i) for i in range(len(doc))] + def get_labels(self,doc): + return [tag for (token,postag,tag) in doc] + def get_model(self): + return self.crf diff --git a/pythainlp/number/__init__.py b/pythainlp/number/__init__.py index b8d449afd..d17b86b0e 100644 --- a/pythainlp/number/__init__.py +++ b/pythainlp/number/__init__.py @@ -21,90 +21,113 @@ d1 = 0 #เลขไทยสู่เลข def thai_num_to_num(text): - """รับค่า ''str'' คืนค่า ''str'' เป็นเลขไทยสู่เลข""" - thaitonum = dict((x[2], x[1]) for x in p[1:]) - return thaitonum[text] -#เลขไทยสู่ข้อความ + """ + :param str text: Thai number characters such as '๑', '๒', '๓' + :return: universal numbers such as '1', '2', '3' + """ + thaitonum = dict((x[2], x[1]) for x in p[1:]) + return thaitonum[text] + def thai_num_to_text(text): - """รับค่า ''str'' คืนค่า ''str'' เป็นเลขไทยสู่ข้อความ""" - thaitonum = dict((x[2], x[0]) for x in p[1:]) - return thaitonum[text] -#เลขสู่เลขไทย + """ + :param str text: Thai number characters such as '๑', '๒', '๓' + :return: Thai numbers, spelled out in Thai + """ + thaitonum = dict((x[2], x[0]) for x in p[1:]) + return thaitonum[text] + def num_to_thai_num(text): - """รับค่า ''str'' คืนค่า ''str'' เป็นเลขสู่เลขไทย""" - thaitonum = dict((x[1], x[2]) for x in p[1:]) - return thaitonum[text] -#เลขสู่ข้อความ + """ + :param text: universal numbers such as '1', '2', '3' + :return: Thai number characters such as '๑', '๒', '๓' + """ + thaitonum = dict((x[1], x[2]) for x in p[1:]) + return thaitonum[text] + def num_to_text(text): - """รับค่า ''str'' คืนค่า ''str'' เป็นเลขสู่ข้อความ""" - thaitonum = dict((x[1], x[0]) for x in p[1:]) - return thaitonum[text] -#ข้อความสู่เลข + """ + :param text: universal numbers such as '1', '2', '3' + :return: Thai numbers, spelled out in Thai + """ + thaitonum = dict((x[1], x[0]) for x in p[1:]) + return thaitonum[text] + def text_to_num(text): - """รับค่า ''str'' คืนค่า ''str'' เป็นข้อความสู่เลข""" - thaitonum = dict((x[0], x[1]) for x in p[1:]) - return thaitonum[text] -#ข้อความสู่เลขไทย + """ + :param text: Thai numbers, spelled out in Thai + :return: universal numbers such as '1', '2', '3' + """ + thaitonum = dict((x[0], x[1]) for x in p[1:]) + return thaitonum[text] + def text_to_thai_num(text): - """รับค่า ''str'' คืนค่า ''str'' เป็นข้อความสู่เลขไทย""" - thaitonum = dict((x[0], x[2]) for x in p[1:]) - return thaitonum[text] + """ + :param text: Thai numbers, spelled out in Thai + :return: Thai numbers such as '๑', '๒', '๓' + """ + thaitonum = dict((x[0], x[2]) for x in p[1:]) + return thaitonum[text] + def number_format(num, places=0): return '{:20,.2f}'.format(num) # fork by http://justmindthought.blogspot.com/2012/12/code-php.html + def numtowords(amount_number): - amount_number = number_format(amount_number, 2).replace(" ","") - pt = amount_number.find(".") - number,fraction = "","" - amount_number1 = amount_number.split('.') - if (pt == False): - number = amount_number - else: - amount_number = amount_number.split('.') - number = amount_number[0] - fraction = int(amount_number1[1]) - ret = "" - number=ast.literal_eval(number.replace(",","")) - baht = ReadNumber(number) - if (baht != ""): - ret += baht + "บาท" - satang = ReadNumber(fraction) - if (satang != ""): - ret += satang + "สตางค์" - else: - ret += "ถ้วน" - return ret + amount_number = number_format(amount_number, 2).replace(" ","") + pt = amount_number.find(".") + number,fraction = "","" + amount_number1 = amount_number.split('.') + if (pt == False): + number = amount_number + else: + amount_number = amount_number.split('.') + number = amount_number[0] + fraction = int(amount_number1[1]) + ret = "" + number=ast.literal_eval(number.replace(",","")) + baht = readnumber(number) + if (baht != ""): + ret += baht + "บาท" + satang = readnumber(fraction) + if (satang != ""): + ret += satang + "สตางค์" + else: + ret += "ถ้วน" + return ret + +def readnumber(number): + """ + :param float number: a float number (with decimals) indicating a quantity + :return: a text that indicates the full amount in word form, properly ending each digit with the right term. + """ + position_call = ["แสน", "หมื่น", "พัน", "ร้อย", "สิบ", ""] + number_call = ["", "หนึ่ง", "สอง", "สาม","สี่", "ห้า", "หก", "เจ็ด", "แปด", "เก้า"] + number = number + ret = "" + if (number == 0): return ret + if (number > 1000000): + ret += readnumber(int(number / 1000000)) + "ล้าน" + number = int(math.fmod(number, 1000000)) + divider = 100000 + pos = 0 + while(number > 0): + d=int(number/divider) + if (divider == 10) and (d == 2): + ret += "ยี่" + elif (divider == 10) and (d == 1): + ret += "" + elif ((divider == 1) and (d == 1) and (ret != "")): + ret += "เอ็ด" + else: + ret += number_call[d] + if d: + ret += position_call[pos] + else: + ret += "" + number=number % divider + divider=divider / 10 + pos += 1 + return ret -#อ่านจำนวนตัวเลขภาษาไทย -def ReadNumber(number): - """อ่านจำนวนตัวเลขภาษาไทย รับค่าเป็น ''float'' คืนค่าเป็น ''str''""" - position_call = ["แสน", "หมื่น", "พัน", "ร้อย", "สิบ", ""] - number_call = ["", "หนึ่ง", "สอง", "สาม","สี่", "ห้า", "หก", "เจ็ด", "แปด", "เก้า"] - number = number - ret = "" - if (number == 0): return ret - if (number > 1000000): - ret += ReadNumber(int(number / 1000000)) + "ล้าน" - number = int(math.fmod(number, 1000000)) - divider = 100000 - pos = 0 - while(number > 0): - d=int(number/divider) - if (divider == 10) and (d == 2): - ret += "ยี่" - elif (divider == 10) and (d == 1): - ret += "" - elif ((divider == 1) and (d == 1) and (ret != "")): - ret += "เอ็ด" - else: - ret += number_call[d] - if d: - ret += position_call[pos] - else: - ret += "" - number=number % divider - divider=divider / 10 - pos += 1 - return ret if __name__ == "__main__": - print(numtowords(4000.0)) \ No newline at end of file + print(numtowords(4000.0)) diff --git a/pythainlp/romanization/__init__.py b/pythainlp/romanization/__init__.py index c9a949772..34593bbd9 100644 --- a/pythainlp/romanization/__init__.py +++ b/pythainlp/romanization/__init__.py @@ -3,14 +3,11 @@ from pythainlp.tokenize import word_tokenize # ถอดเสียงภาษาไทยเป็น Latin def romanization(data,engine='royin'): - """เป็นคำสั่ง ถอดเสียงภาษาไทยเป็นอังกฤษ - romanization(data,engine='royin') - มี 2 engine ดังนี้ - - pyicu ส่งค่า Latin - - royin ใช้หลักเกณฑ์การถอดอักษรไทยเป็นอักษรโรมัน ฉบับราชบัณฑิตยสถาน - data : - รับค่า ''str'' ข้อความ - คืนค่าเป็น ''str'' ข้อความ""" + """ + :param str data: Thai text to be romanized + :param str engine: choose between 'royin' , 'pyicu' and 'thai2rom'. 'royin' will romanize according to the standard of Thai Royal Institute. 'pyicu' will romanize according to the Internaitonal Phonetic Alphabet. 'thai2rom' is deep learning thai romanization. + :return: English (more or less) text that spells out how the Thai text should read. + """ word_list=word_tokenize(data) listword=[] i=0 @@ -18,6 +15,12 @@ def romanization(data,engine='royin'): from .royin import romanization elif engine=='pyicu': from .pyicu import romanization + elif engine=='thai2rom': + from pythainlp.romanization.thai2rom import thai2rom + thai=thai2rom() + return thai.romanization(data) + else: + raise Exception("error no have engine.") while i self.max_decoder_seq_length): + self.stop_condition = True + self.target_seq = np.zeros((1, 1, self.num_decoder_tokens)) + self.target_seq[0, 0, self.sampled_token_index] = 1. + self.states_value = [self.h, self.c] + return self.decoded_sentence + def encode_input(self,name): + self.test_input = np.zeros((1, self.max_encoder_seq_length, self.num_encoder_tokens),dtype='float32') + for t, char in enumerate(name): + self.test_input[0, t, self.input_token_index[char]] = 1. + return self.test_input + def romanization(self,text): + ''' + :param str text: Thai text to be romanized + :return: English (more or less) text that spells out how the Thai text should read. + ''' + return self.decode_sequence(self.encode_input(text)) diff --git a/pythainlp/sentiment/.ipynb_checkpoints/__init__-checkpoint.py b/pythainlp/sentiment/.ipynb_checkpoints/__init__-checkpoint.py new file mode 100644 index 000000000..77275d0db --- /dev/null +++ b/pythainlp/sentiment/.ipynb_checkpoints/__init__-checkpoint.py @@ -0,0 +1,40 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,unicode_literals,print_function +import six +import sys +if six.PY2: + print("Thai sentiment in pythainlp. Not support python 2.7") + sys.exit(0) +import pythainlp +from pythainlp.corpus import stopwords +import os +from pythainlp.tokenize import word_tokenize +try: + import dill +except ImportError: + import pip + pip.main(['install','dill']) + try: + import dill + except ImportError: + print("Error ! using 'pip install dill'") + sys.exit(0) + +templates_dir = os.path.join(os.path.dirname(pythainlp.__file__), 'sentiment') +def sentiment(text): + """ + sentiment ภาษาไทย + ใช้ข้อมูลจาก https://github.com/wannaphongcom/lexicon-thai/tree/master/ข้อความ/ + รับค่าสตริง str คืนค่า pos , neg""" + with open(os.path.join(templates_dir, 'vocabulary.data'), 'rb') as in_strm: + vocabulary = dill.load(in_strm) + in_strm.close() + with open(os.path.join(templates_dir, 'sentiment.data'), 'rb') as in_strm: + classifier = dill.load(in_strm) + in_strm.close() + text=set(word_tokenize(text))-set(stopwords.words('thai')) + featurized_test_sentence = {i:(i in text) for i in vocabulary} + return classifier.classify(featurized_test_sentence) +if __name__ == '__main__': + d="เสียใจแย่มากเลย" + print(sentiment(d)) \ No newline at end of file diff --git a/pythainlp/sentiment/.ipynb_checkpoints/build_pythainlp-checkpoint.tool b/pythainlp/sentiment/.ipynb_checkpoints/build_pythainlp-checkpoint.tool new file mode 100644 index 000000000..b7a7f7107 --- /dev/null +++ b/pythainlp/sentiment/.ipynb_checkpoints/build_pythainlp-checkpoint.tool @@ -0,0 +1,48 @@ +# -*- coding: utf-8 -*- +# เครื่องมือในการ build sentiment เพื่อใช้ใน pythainlp +# เวชั่น 0.3 +# 2018/01/18 +# เขียนโดย นาย วรรณพงษ์ ภัททิยไพบูลย์ +# cc-by 3.0 Thai Sentiment Text https://github.com/wannaphongcom/lexicon-thai/tree/master/ข้อความ/ +# อ่านบทความได้ที่ https://python3.wannaphong.com/2017/02/ทำ-sentiment-analysis-ภาษาไทยใน-python.html +from nltk import NaiveBayesClassifier as nbc +import dill +from pythainlp.tokenize import word_tokenize +from pythainlp.corpus import stopwords +import codecs +from itertools import chain +thaistopwords = stopwords.words('thai') +# pos.txt +with codecs.open('pos.txt', 'r', "utf-8") as f: + lines = f.readlines() +listpos=[x for x in [e.strip() for e in lines] if x not in thaistopwords] +del lines +f.close() # ปิดไฟล์ +# neg.txt +with codecs.open('neg.txt', 'r', "utf-8") as f: + lines = f.readlines() +listneg=[x for x in [e.strip() for e in lines] if x not in thaistopwords] +f.close() # ปิดไฟล์ +print(1) +pos1=['pos']*len(listpos) +neg1=['neg']*len(listneg) +print(2) +training_data = list(zip(listpos,pos1)) + list(zip(listneg,neg1)) +print(3) +#vocabulary = set(chain(*[(set(word_tokenize(i[0]))-set(stopwords.words('thai'))) for i in training_data])) +#vocabulary = set(chain(*[x for x in a if x not in [list(set(word_tokenize(i[0]))) for i in training_data]])) +vocabulary = set(chain(*[word_tokenize(i[0]) for i in training_data])) +#print(vocabulary) +print(3.1) +feature_set = [({i:(i in word_tokenize(sentence)) for i in vocabulary},tag) for sentence, tag in training_data] +#print(feature_set) +print(4) +classifier = nbc.train(feature_set) +print(5) +with open('vocabulary.data', 'wb') as out_strm: + dill.dump(vocabulary,out_strm) +out_strm.close() +with open('sentiment.data', 'wb') as out_strm: + dill.dump(classifier,out_strm) +out_strm.close() +print('OK') \ No newline at end of file diff --git a/pythainlp/sentiment/.ipynb_checkpoints/ulmfit_sent-checkpoint.py b/pythainlp/sentiment/.ipynb_checkpoints/ulmfit_sent-checkpoint.py new file mode 100644 index 000000000..aacd9e844 --- /dev/null +++ b/pythainlp/sentiment/.ipynb_checkpoints/ulmfit_sent-checkpoint.py @@ -0,0 +1,89 @@ +# -*- coding: utf-8 -*- +''' +Code by https://github.com/cstorm125/thai2vec/tree/master/notebook +''' +from __future__ import absolute_import,unicode_literals +import os +import sys +from collections import defaultdict + +#numpy and dill +try: + import numpy as np + import dill as pickle +except ImportError: + import pip + pip.main(['install','numpy','dill']) + try: + import numpy as np + import dill as pickle + except ImportError: + print("Error installing using 'pip install numpy dill'") + sys.exit(0) + +#import torch +try: + import torch +except ImportError: + print('PyTorch required. See https://pytorch.org/.') +import torch +from torch.autograd import Variable +from torch import LongTensor + +#import fastai for multiBatchRNN +try: + from fastai.text import * +except ImportError: + print( + """ + fastai required for multiBatchRNN. + Run 'pip install https://github.com/fastai/fastai/archive/master.zip' + """) + +from pythainlp.tokenize import word_tokenize +from pythainlp.corpus import get_file +from pythainlp.corpus import download + +MODEL_NAME = 'sent_model' +ITOS_NAME = 'itos_sent' + +#download pretrained model +def get_path(fname): + path = get_file(fname) + if path==None: + download(fname) + path = get_file(fname) + return(path) + +#load model +m = torch.load(get_path(MODEL_NAME)) +m.eval() +#load itos and stoi +itos = pickle.load(open(get_path(ITOS_NAME),'rb')) +stoi = defaultdict(lambda:0, {v:k for k,v in enumerate(itos)}) + + +#get sentiment; 1 for positive and 0 for negative +#or score if specified return_score=True +softmax = lambda x : np.exp(x)/np.sum(np.exp(x)) +def get_sentiment(ss,return_score=False): + s = word_tokenize(ss) + t = LongTensor([stoi[i] for i in s]).view(-1,1).cpu() + t = Variable(t,volatile=False) + m.reset() + pred,*_ = m(t) + result = pred.data.cpu().numpy().reshape(-1) + if return_score: + return(softmax(result)) + else: + return(np.argmax(result)) + +def about(): + return ''' + Sentiment Analyzer based on thai2vec + Data is from various online reviews including but not limited to JagerV3 and Wongnai Challenge. + 89% accuracy based on 15% validation set compared to 72% of fastText and 52% most-frequent-class baseline. + + Development : Charin Polpanumas + GitHub : https://github.com/cstorm125/thai2vec + ''' \ No newline at end of file diff --git a/pythainlp/sentiment/__init__.py b/pythainlp/sentiment/__init__.py index 77275d0db..af7995af2 100644 --- a/pythainlp/sentiment/__init__.py +++ b/pythainlp/sentiment/__init__.py @@ -1,40 +1,50 @@ # -*- coding: utf-8 -*- from __future__ import absolute_import,unicode_literals,print_function -import six -import sys -if six.PY2: - print("Thai sentiment in pythainlp. Not support python 2.7") - sys.exit(0) import pythainlp from pythainlp.corpus import stopwords import os from pythainlp.tokenize import word_tokenize -try: - import dill -except ImportError: - import pip - pip.main(['install','dill']) - try: - import dill - except ImportError: - print("Error ! using 'pip install dill'") - sys.exit(0) +import dill templates_dir = os.path.join(os.path.dirname(pythainlp.__file__), 'sentiment') -def sentiment(text): +def sentiment(text, engine='old'): """ - sentiment ภาษาไทย - ใช้ข้อมูลจาก https://github.com/wannaphongcom/lexicon-thai/tree/master/ข้อความ/ - รับค่าสตริง str คืนค่า pos , neg""" - with open(os.path.join(templates_dir, 'vocabulary.data'), 'rb') as in_strm: - vocabulary = dill.load(in_strm) - in_strm.close() - with open(os.path.join(templates_dir, 'sentiment.data'), 'rb') as in_strm: - classifier = dill.load(in_strm) - in_strm.close() - text=set(word_tokenize(text))-set(stopwords.words('thai')) - featurized_test_sentence = {i:(i in text) for i in vocabulary} - return classifier.classify(featurized_test_sentence) + :param str text: thai text + :param str engine: sentiment analysis engine (old or ulmfit) + :return: pos or neg + + **Example**:: + >>> from pythainlp.sentiment import sentiment + >>> text="วันนี้อากาศดีจัง" + >>> sentiment(text) + 'pos' + >>> sentiment(text,'ulmfit') + 'pos' + >>> text="วันนี้อารมณ์เสียมาก" + >>> sentiment(text) + 'neg' + >>> sentiment(text,'ulmfit') + 'neg' + """ + if engine=='old': + with open(os.path.join(templates_dir, 'vocabulary.data'), 'rb') as in_strm: + vocabulary = dill.load(in_strm) + with open(os.path.join(templates_dir, 'sentiment.data'), 'rb') as in_strm: + classifier = dill.load(in_strm) + text=set(word_tokenize(text))-set(stopwords.words('thai')) + featurized_test_sentence = {i:(i in text) for i in vocabulary} + return classifier.classify(featurized_test_sentence) + elif engine=='ulmfit': + from pythainlp.sentiment import ulmfit_sent + tag=ulmfit_sent.get_sentiment(text) + sa="" + if tag==0: + sa="neg" + else: + sa="pos" + return sa + else: + raise Exception("error no have engine.") if __name__ == '__main__': d="เสียใจแย่มากเลย" - print(sentiment(d)) \ No newline at end of file + print(sentiment(d)) diff --git a/pythainlp/sentiment/ulmfit_sent.py b/pythainlp/sentiment/ulmfit_sent.py new file mode 100644 index 000000000..85bd7b790 --- /dev/null +++ b/pythainlp/sentiment/ulmfit_sent.py @@ -0,0 +1,90 @@ +# -*- coding: utf-8 -*- +''' +Code by https://github.com/cstorm125/thai2vec/tree/master/notebook +''' +from __future__ import absolute_import,unicode_literals +import os +import sys +from collections import defaultdict + +#numpy and dill +try: + import numpy as np + import dill as pickle +except ImportError: + from pythainlp.tools import install_package + install_package('numpy') + install_package('dill') + try: + import numpy as np + import dill as pickle + except ImportError: + print("Error installing using 'pip install numpy dill'") + sys.exit(0) + +#import torch +try: + import torch +except ImportError: + print('PyTorch required. See https://pytorch.org/.') +import torch +from torch.autograd import Variable +from torch import LongTensor + +#import fastai for multiBatchRNN +try: + from fastai.text import * +except ImportError: + print( + """ + fastai required for multiBatchRNN. + Run 'pip install https://github.com/fastai/fastai/archive/master.zip' + """) + +from pythainlp.tokenize import word_tokenize +from pythainlp.corpus import get_file +from pythainlp.corpus import download + +MODEL_NAME = 'sent_model' +ITOS_NAME = 'itos_sent' + +#download pretrained model +def get_path(fname): + path = get_file(fname) + if path==None: + download(fname) + path = get_file(fname) + return(path) + +#load model +m = torch.load(get_path(MODEL_NAME)) +m.eval() +#load itos and stoi +itos = pickle.load(open(get_path(ITOS_NAME),'rb')) +stoi = defaultdict(lambda:0, {v:k for k,v in enumerate(itos)}) + + +#get sentiment; 1 for positive and 0 for negative +#or score if specified return_score=True +softmax = lambda x : np.exp(x)/np.sum(np.exp(x)) +def get_sentiment(ss,return_score=False): + s = word_tokenize(ss) + t = LongTensor([stoi[i] for i in s]).view(-1,1).cpu() + t = Variable(t,volatile=False) + m.reset() + pred,*_ = m(t) + result = pred.data.cpu().numpy().reshape(-1) + if return_score: + return(softmax(result)) + else: + return(np.argmax(result)) + +def about(): + return ''' + Sentiment Analyzer based on thai2vec + Data is from various online reviews including but not limited to JagerV3 and Wongnai Challenge. + 89% accuracy based on 15% validation set compared to 72% of fastText and 52% most-frequent-class baseline. + + Development : Charin Polpanumas + GitHub : https://github.com/cstorm125/thai2vec + ''' \ No newline at end of file diff --git a/pythainlp/soundex.py b/pythainlp/soundex.py index 75f89af7c..c7d5f3a69 100644 --- a/pythainlp/soundex.py +++ b/pythainlp/soundex.py @@ -9,8 +9,10 @@ import re def LK82(s): ''' - LK82 - กฎการเข้ารหัสซาวน์เด็กซ์ของ วิชิตหล่อจีระชุณห์กุล และ เจริญ คุวินทร์พันธุ์ - LK82(str) + LK82 - It's a thai soundex rule. + + :param str s: thai word + :return: LK82 soundex ''' t1 = str.maketrans("กขฃคฅฆงจฉชฌซศษสญยฎดฏตณนฐฑฒถทธบปผพภฝฟมรลฬฤฦวหฮอ","กกกกกกงจชชชซซซซยยดดตตนนททททททบปพพพฟฟมรรรรรวหหอ") t2 = str.maketrans("กขฃคฅฆงจฉชซฌฎฏฐฑฒดตถทธศษสญณนรลฬฤฦบปพฟภผฝมำยวไใหฮาๅึืเแโุูอ","1111112333333333333333333444444445555555667777889AAABCDEEF") @@ -57,6 +59,12 @@ def LK82(s): # 14. เติมศูนย์ให้ครบ ถ้าเกินก็ตัด return ("".join(res2)+"0000")[:5] def Udom83(s): + ''' + Udom83 - It's a thai soundex rule. + + :param str s: thai word + :return: LK82 soundex + ''' tu1 = str.maketrans("กขฃคฅฆงจฉชฌซศษสฎดฏตฐฑฒถทธณนบปผพภฝฟมญยรลฬฤฦวอหฮ" ,"กขขขขขงจชชชสสสสดดตตททททททนนบปพพพฟฟมยยรรรรรวอฮฮ") tu2 = str.maketrans("มวำกขฃคฅฆงยญณนฎฏดตศษสบปพภผฝฟหอฮจฉชซฌฐฑฒถทธรฤลฦ","0001111112233344444445555666666777778888889999") s = re.sub('รร([เ-ไ])', 'ัน\\1', s) # 4. diff --git a/pythainlp/spell/__init__.py b/pythainlp/spell/__init__.py index ed0cdf1c0..df503e3d7 100644 --- a/pythainlp/spell/__init__.py +++ b/pythainlp/spell/__init__.py @@ -1,12 +1,13 @@ # -*- coding: utf-8 -*- from __future__ import absolute_import,unicode_literals def spell(word,engine='pn'): - ''' - คำสั่งเช็คคำผิด spell(word,engine='pn') - engine ที่รองรับ - - pn พัฒนามาจาก Peter Norvig (ค่าเริ่มต้น) - - hunspell ใช้ hunspell (ไม่รองรับ Python 2.7) - ''' + """ + :param str word: the word to check spelling + :param str engine: + * pn - Peter Norvig's algorithm + * hunspell - uses hunspell's algorithm, which should already exist in linux + :return: list word + """ if engine=='pn': from .pn import spell as spell1 elif engine=='hunspell': diff --git a/pythainlp/spell/hunspell.py b/pythainlp/spell/hunspell.py index 586781b72..c940126f5 100644 --- a/pythainlp/spell/hunspell.py +++ b/pythainlp/spell/hunspell.py @@ -2,11 +2,7 @@ from __future__ import absolute_import,print_function,unicode_literals from builtins import * import subprocess -import six import sys -if six.PY2: - print("Not support python 2.7") - sys.exit(0) def spel1(word,lang='th_TH'): """เป็นคำสั่งตรวจคำผิดโดยใช้ hunspell @@ -33,7 +29,7 @@ def spel1(word,lang='th_TH'): del get return getoutput except subprocess.CalledProcessError: - print('plase install hunspell') + print('please install hunspell') return None if __name__ == "__main__": Input = spell("appoe","") @@ -41,4 +37,4 @@ def spel1(word,lang='th_TH'): InputTH = spell("คลินิค","th_TH") print(InputTH) trueth = spell("สี่เหลียม","th_TH") - print(trueth) \ No newline at end of file + print(trueth) diff --git a/pythainlp/summarize/__init__.py b/pythainlp/summarize/__init__.py index 4bac452fc..0c413bd7b 100644 --- a/pythainlp/summarize/__init__.py +++ b/pythainlp/summarize/__init__.py @@ -27,9 +27,9 @@ def _compute_frequencies(self, word_sent): def _rank(self, ranking, n): return nlargest(n, ranking, key=ranking.get) - def summarize(self, text, n): + def summarize(self, text, n,tokenize): sents = sent_tokenize(text) - word_sent = [word_tokenize(s) for s in sents] + word_sent = [word_tokenize(s,tokenize) for s in sents] self._freq = self._compute_frequencies(word_sent) ranking = defaultdict(int) for i, sent in enumerate(word_sent): @@ -38,15 +38,14 @@ def summarize(self, text, n): ranking[i] += self._freq[w] sents_idx = self._rank(ranking,n) return [sents[j] for j in sents_idx] -def summarize_text(text,n,engine='frequency'): +def summarize_text(text,n,engine='frequency',tokenize='newmm'): ''' - คำสั่งสรุปเอกสารภาษาไทย - summarize_text(text,n,engine='frequency') - text เป็นข้อความ - n คือ จำนวนประโยคสรุป - engine ที่รองรับ - - frequency + Thai text summarize. + :param str text: thai text + :param int n: sent number + :param str engine: Thai text summarize engine. + :param str tokenize: thai word tokenize. ''' if engine=='frequency': - data=FrequencySummarizer().summarize(text,n) + data=FrequencySummarizer().summarize(text,n,tokenize) return data diff --git a/pythainlp/tag/__init__.py b/pythainlp/tag/__init__.py index 159c5d69a..c2c45f7c7 100644 --- a/pythainlp/tag/__init__.py +++ b/pythainlp/tag/__init__.py @@ -1,34 +1,43 @@ # -*- coding: utf-8 -*- from __future__ import absolute_import,division,print_function,unicode_literals import sys -def pos_tag(text,engine='old'): - """ - ระบบ postaggers +def pos_tag(list_text,engine='unigram',corpus='orchid'): + """ + Part of Speech tagging function. - pos_tag(text,engine='old') - engine ที่รองรับ - * old เป็น UnigramTagger - * artagger เป็น RDR POS Tagger - """ - if engine=='old': - from .old import tag - elif engine=='artagger': - if sys.version_info < (3,4): - sys.exit('Sorry, Python < 3.4 is not supported') - def tag(text1): - try: - from artagger import Tagger - except ImportError: - import pip - pip.main(['install','https://github.com/franziz/artagger/archive/master.zip']) - try: - from artagger import Tagger - except ImportError: - print("Error ! using 'pip install https://github.com/franziz/artagger/archive/master.zip'") - sys.exit(0) - words = Tagger().tag(' '.join(text1)) - totag=[] - for word in words: - totag.append((word.word, word.tag)) - return totag - return tag(text) + :param list list_text: takes in a list of tokenized words (put differently, a list of string) + :param str engine: + * unigram - unigram tagger + * perceptron - perceptron tagger + * artagger - RDR POS tagger + :param str corpus: + * orchid - annotated Thai academic articles + * pud - Parallel Universal Dependencies (PUD) treebanks + :return: returns a list of labels regarding which part of speech it is + """ + if engine=='old' or engine=='unigram': + from .old import tag + elif engine=='perceptron': + from .perceptron import tag + elif engine=='artagger': + def tag(text1): + try: + from artagger import Tagger + except ImportError: + from pythainlp.tools import install_package + install_package('https://github.com/wannaphongcom/artagger/archive/master.zip') + try: + from artagger import Tagger + except ImportError: + print("Error ! using 'pip install https://github.com/wannaphongcom/artagger/archive/master.zip'") + sys.exit(0) + words = Tagger().tag(' '.join(text1)) + totag=[] + for word in words: + totag.append((word.word, word.tag)) + return totag + return tag(list_text) + return tag(list_text,corpus=corpus) + +def pos_tag_sents(sentences,engine='unigram',corpus='orchid'): + return [pos_tag(i,engine=engine,corpus=corpus) for i in sentences] diff --git a/pythainlp/tag/old.py b/pythainlp/tag/old.py index 6a220faf6..acaf72841 100644 --- a/pythainlp/tag/old.py +++ b/pythainlp/tag/old.py @@ -3,17 +3,26 @@ import codecs import os import json +import pythainlp import nltk.tag -import nltk.data -def data(): - import pythainlp - templates_dir = os.path.join(os.path.dirname(pythainlp.__file__), 'corpus') +import dill +templates_dir = os.path.join(os.path.dirname(pythainlp.__file__), 'corpus') +def orchid_data(): template_file = os.path.join(templates_dir, 'thaipos.json') with codecs.open(template_file,'r',encoding='utf-8-sig') as handle: model = json.load(handle) return model -def tag(text): +def pud_data(): + template_file = os.path.join(templates_dir, 'ud_thai-pud_unigram_tagger.dill') + with open(template_file,'rb') as handle: + model = dill.load(handle) + return model +def tag(text,corpus): """ รับค่าเป็น ''list'' คืนค่าเป็น ''list'' เช่น [('ข้อความ', 'ชนิดคำ')]""" - tagger = nltk.tag.UnigramTagger(model=data())# backoff=default_tagger) - return tagger.tag(text) + if corpus=='orchid': + tagger = nltk.tag.UnigramTagger(model=orchid_data())# backoff=default_tagger) + return tagger.tag(text) + elif corpus=='pud': + tagger = pud_data() + return tagger.tag(text) diff --git a/pythainlp/tag/perceptron.py b/pythainlp/tag/perceptron.py new file mode 100644 index 000000000..16ce35969 --- /dev/null +++ b/pythainlp/tag/perceptron.py @@ -0,0 +1,27 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,division,unicode_literals +import sys +import os +import pythainlp +import nltk.tag +import dill +templates_dir = os.path.join(os.path.dirname(pythainlp.__file__), 'corpus') +def orchid_data(): + template_file = os.path.join(templates_dir, 'pt_tagger_1.dill') + with open(template_file,'rb') as handle: + model = dill.load(handle) + return model +def pud_data(): + template_file = os.path.join(templates_dir, 'ud_thai-pud_pt_tagger.dill') + with open(template_file,'rb') as handle: + model = dill.load(handle) + return model +def tag(text,corpus): + """ + รับค่าเป็น ''list'' คืนค่าเป็น ''list'' เช่น [('ข้อความ', 'ชนิดคำ')]""" + if corpus=='orchid': + tagger = orchid_data() + return tagger.tag(text) + elif corpus=='pud': + tagger = pud_data() + return tagger.tag(text) diff --git a/pythainlp/test/__init__.py b/pythainlp/test/__init__.py index 857a25531..970ad3c36 100644 --- a/pythainlp/test/__init__.py +++ b/pythainlp/test/__init__.py @@ -4,11 +4,11 @@ from collections import Counter from pythainlp.corpus import alphabet,wordnet,country,tone,provinces,stopwords,newthaiword,thaiword from pythainlp.keywords import * -from pythainlp.tokenize import word_tokenize,tcc,etcc,isthai,WhitespaceTokenizer,syllable_tokenize +from pythainlp.tokenize import word_tokenize,tcc,etcc,isthai,syllable_tokenize from pythainlp.rank import rank from pythainlp.change import texttothai,texttoeng from pythainlp.number import numtowords -from pythainlp.tag import pos_tag +from pythainlp.tag import pos_tag,pos_tag_sents from pythainlp.romanization import romanization from pythainlp.date import now,reign_year_to_ad from pythainlp.soundex import LK82,Udom83 @@ -16,23 +16,19 @@ from pythainlp.spell import spell from collections import namedtuple from pythainlp.collation import collation -from pythainlp.util import normalize +from pythainlp.util import normalize,listtext_num2num from pythainlp.summarize import summarize_text +from pythainlp.ner import thainer class TestUM(unittest.TestCase): """ - ระบบทดสอบการทำงานของโค้ดของ PyThaiNLP 1.6 + ระบบทดสอบการทำงานของโค้ดของ PyThaiNLP 1.7 """ def test_segment(self): self.assertEqual(word_tokenize('ฉันรักภาษาไทยเพราะฉันเป็นคนไทย'),[u'ฉัน', u'รัก', u'ภาษาไทย', u'เพราะ', u'ฉัน', u'เป็น', u'คนไทย']) def test_syllable_tokenize(self): self.assertEqual(syllable_tokenize("สวัสดีชาวโลก"),[u'สวัส', u'ดี', u'ชาว', u'โลก']) - def test_syllable_deeplearning(self): - if sys.version_info >= (3,4): - self.assertEqual(word_tokenize('ฉันรักภาษาไทยเพราะฉันเป็นคนไทย',engine='cutkum'),[u'ฉัน', u'รัก', u'ภาษา', u'ไทย', u'เพราะ', u'ฉัน', u'เป็น', u'คน', u'ไทย']) def test_segment_icu(self): self.assertEqual(word_tokenize('ฉันรักภาษาไทยเพราะฉันเป็นคนไทย',engine='icu'),[u'ฉัน', u'รัก', u'ภาษา', u'ไทย', u'เพราะ', u'ฉัน', u'เป็น', u'คน', u'ไทย']) - def test_segment_dict(self): - self.assertEqual(word_tokenize('ฉันรักภาษาไทยเพราะฉันเป็นคนไทย',engine='dict'),[u'ฉัน', u'รัก', u'ภาษาไทย', u'เพราะ', u'ฉัน', u'เป็น', u'คนไทย']) def test_segment_mm(self): self.assertEqual(word_tokenize('ฉันรักภาษาไทยเพราะฉันเป็นคนไทย',engine='mm'),[u'ฉัน', u'รัก', u'ภาษาไทย', u'เพราะ', u'ฉัน', u'เป็น', u'คนไทย']) def test_segment_newmm(self): @@ -40,10 +36,11 @@ def test_segment_newmm(self): self.assertEqual(word_tokenize('สวัสดีครับ สบายดีไหมครับ',engine='newmm'),[u'สวัสดี', u'ครับ', u' ', u'สบายดี', u'ไหม', u'ครับ']) self.assertEqual(word_tokenize('จุ๋มง่วงนอนยัง',engine='newmm'),[u'จุ๋ม', u'ง่วงนอน', u'ยัง']) self.assertEqual(word_tokenize('จุ๋มง่วง',engine='newmm'),[u'จุ๋ม', u'ง่วง']) + self.assertEqual(word_tokenize('จุ๋ม ง่วง',engine='newmm',whitespaces=False),[u'จุ๋ม', u'ง่วง']) def test_segment_longest_matching(self): self.assertEqual(word_tokenize('ฉันรักภาษาไทยเพราะฉันเป็นคนไทย',engine='longest-matching'),[u'ฉัน', u'รัก', u'ภาษาไทย', u'เพราะ', u'ฉัน', u'เป็น', u'คนไทย']) def test_segment_Wordcut(self): - if sys.version_info >= (3,4): + if sys.version_info >= (3,4) and sys.platform!="win32" and sys.platform!="win64": self.assertEqual(word_tokenize('ฉันรักภาษาไทยเพราะฉันเป็นคนไทย',engine='wordcutpy'),[u'ฉัน', u'รัก', u'ภาษา', u'ไทย', u'เพราะ', u'ฉัน', u'เป็น', u'คน', u'ไทย']) def test_rank(self): self.assertEqual(rank(["แมว","คน","แมว"]),Counter({'แมว': 2, 'คน': 1})) @@ -65,8 +62,8 @@ def test_tcc(self): self.assertEqual(tcc.tcc('ประเทศไทย'),'ป/ระ/เท/ศ/ไท/ย') def test_isthai(self): self.assertEqual(isthai('ประเทศไทย'),{'thai': 100.0}) - def test_WhitespaceTokenizer(self): - self.assertEqual(WhitespaceTokenizer("1 2 3"),['1', '2', '3']) + # def test_WhitespaceTokenizer(self): + # self.assertEqual(WhitespaceTokenizer("1 2 3"),['1', '2', '3']) def test_etcc(self): self.assertEqual(etcc.etcc('คืนความสุข'),'/คืน/ความสุข') def test_lk82(self): @@ -97,11 +94,19 @@ def test_collation(self): self.assertEqual(collation(['ไก่', 'เป็ด', 'หมู', 'วัว']),[u'ไก่', u'เป็ด', u'วัว', u'หมู']) def test_normalize(self): self.assertEqual(normalize("เเปลก"),"แปลก") + def test_listtext_num2num(self): + if sys.version_info >= (3,4): + self.assertEqual(listtext_num2num([u'หก',u'ล้าน',u'หกแสน',u'หกหมื่น',u'หกพัน',u'หกร้อย',u'หกสิบ',u'หก']),6666666) def test_keywords(self): self.assertEqual(find_keyword(word_tokenize("แมวกินปลาอร่อยรู้ไหมว่าแมวเป็นแมวรู้ไหมนะแมว",engine='newmm')),{u'แมว': 4}) def test_tag(self): self.assertEqual(pos_tag(word_tokenize("คุณกำลังประชุม"),engine='old'),[('คุณ', 'PPRS'), ('กำลัง', 'XVBM'), ('ประชุม', 'VACT')]) + self.assertEqual(pos_tag_sents([["ผม","กิน","ข้าว"],["แมว","วิ่ง"]]),[[('ผม', 'PPRS'), ('กิน', 'VACT'), ('ข้าว', 'NCMN')], [('แมว', 'NCMN'), ('วิ่ง', 'VACT')]]) if sys.version_info >= (3,4): self.assertEqual(str(type(pos_tag(word_tokenize("ผมรักคุณ"),engine='artagger'))),"") + def test_ner(self): + ner=thainer() + self.assertEqual(ner.get_ner("แมวทำอะไรตอนห้าโมงเช้า"),[('แมว', 'NCMN', 'O'),('ทำ', 'VACT', 'O'),('อะไร', 'PNTR', 'O'),('ตอน', 'NCMN', 'O'),('ห้า', 'VSTA', 'B-TIME'),('โมง', 'NCMN', 'I-TIME'),('เช้า', 'ADVN', 'I-TIME')]) + self.assertEqual(ner.get_ner("แมวทำอะไรตอนห้าโมงเช้า",postag=False),[('แมว', 'O'),('ทำ', 'O'),('อะไร', 'O'),('ตอน', 'O'),('ห้า', 'B-TIME'),('โมง', 'I-TIME'),('เช้า', 'I-TIME')]) if __name__ == '__main__': unittest.main() diff --git a/pythainlp/tokenize/.ipynb_checkpoints/__init__-checkpoint.py b/pythainlp/tokenize/.ipynb_checkpoints/__init__-checkpoint.py new file mode 100644 index 000000000..c3d3f3731 --- /dev/null +++ b/pythainlp/tokenize/.ipynb_checkpoints/__init__-checkpoint.py @@ -0,0 +1,227 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,unicode_literals +import nltk +import re +import codecs +from six.moves import zip +from pythainlp.corpus.thaisyllable import get_data +from pythainlp.corpus.thaiword import get_data as get_dict +from marisa_trie import Trie + +DEFAULT_DICT_TRIE = Trie(get_dict()) + +def dict_word_tokenize(text, custom_dict_trie, engine='newmm'): + ''' + dict_word_tokenize(text,file,engine) + เป็นคำสั่งสำหรับตัดคำโดยใช้ข้อมูลที่ผู้ใช้กำหนด + text คือ ข้อความที่ต้องการตัดคำ + custom_dict_trie คือ trie ที่สร้างจาก create_custom_dict_trie + engine คือ เครื่องมือตัดคำ + - newmm ตัดคำด้วย newmm + - wordcutpy ใช้ wordcutpy (https://github.com/veer66/wordcutpy) ในการตัดคำ + - mm ตัดคำด้วย mm + - longest-matching ตัดคำโดยใช้ longest matching + ''' + + if engine=="newmm": + from .newmm import mmcut as segment + elif engine=="mm": + from .mm import segment + elif engine=='longest-matching': + from .longest import segment + elif engine=='wordcutpy': + from .wordcutpy import segment + return segment(text, custom_dict_trie.keys()) + + return segment(text, custom_dict_trie) + +def word_tokenize(text, engine='newmm',whitespaces=True): + """ + ระบบตัดคำภาษาไทย + + word_tokenize(text,engine='newmm') + text คือ ข้อความในรูปแบบ str + engine มี + - newmm - ใช้ Maximum Matching algorithm ในการตัดคำภาษาไทย โค้ดชุดใหม่ (ค่าเริ่มต้น) + - icu - engine ตัวดั้งเดิมของ PyThaiNLP (ความแม่นยำต่ำ) + - dict - ใช้ dicu ในการตัดคำไทย จะคืนค่า False หากไม่สามารถตัดคำไทย + - longest-matching ใช้ Longest matching ในการตัดคำ + - mm ใช้ Maximum Matching algorithm - โค้ดชุดเก่า + - pylexto ใช้ LexTo ในการตัดคำ + - deepcut ใช้ Deep Neural Network ในการตัดคำภาษาไทย + - wordcutpy ใช้ wordcutpy (https://github.com/veer66/wordcutpy) ในการตัดคำ + """ + + if engine=='icu': + ''' + ตัดคำภาษาไทยโดยใช้ icu ในการตัดคำ + คำเตือน !!! \n คำสั่ง word_tokenize(text) ใน PyThaiNLP 1.6 + ค่าเริ่มต้นจะเปลี่ยนจาก icu ไปเป็น newmm''' + from .pyicu import segment + elif engine=='dict': + ''' + ใช้ dicu ในการตัดคำไทย + จะคืนค่า False หากไม่สามารถตัดคำไทย + ''' + from .dictsegment import segment + elif engine=='mm': + ''' + ใช้ Maximum Matching algorithm - โค้ดชุดเก่า + ''' + from .mm import segment + elif engine=='newmm': + ''' + ใช้ Maximum Matching algorithm ในการตัดคำภาษาไทย โค้ดชุดใหม่ + ''' + from .newmm import mmcut as segment + elif engine=='longest-matching': + ''' + ใช้ Longest matching ในการตัดคำ + ''' + from .longest import segment + elif engine=='pylexto': + ''' + ใช้ LexTo ในการตัดคำ + ''' + from .pylexto import segment + elif engine=='deepcut': + ''' + ใช้ Deep Neural Network ในการตัดคำภาษาไทย + ''' + from .deepcut import segment + elif engine=='cutkum': + ''' + ใช้ Deep Neural Network ในการตัดคำภาษาไทย (https://github.com/pucktada/cutkum) + ''' + from .cutkum import segment + elif engine=='wordcutpy': + ''' + wordcutpy ใช้ wordcutpy (https://github.com/veer66/wordcutpy) ในการตัดคำ + ''' + from .wordcutpy import segment + else: + raise Exception("error no have engine.") + if whitespaces==False: + return [i.strip(' ') for i in segment(text) if i.strip(' ')!=''] + return segment(text) + +def sent_tokenize(text,engine='whitespace+newline'): + ''' + sent_tokenize(text,engine='whitespace+newline') + ตัดประโยคเบื้องต้น โดยการแบ่งด้วยช่องว่าง + ''' + if engine=='whitespace': + data=nltk.tokenize.WhitespaceTokenizer().tokenize(text) + elif engine=='whitespace+newline': + data=re.sub(r'\n+|\s+','|',text,re.U).split('|') + return data +def wordpunct_tokenize(text): + ''' + wordpunct_tokenize(text) + It is nltk.tokenize.wordpunct_tokenize(text). + ''' + return nltk.tokenize.wordpunct_tokenize(text) +def WhitespaceTokenizer(text): + return nltk.tokenize.WhitespaceTokenizer().tokenize(text) +def isthai(text,check_all=False): + """ + สำหรับเช็คว่าเป็นตัวอักษรภาษาไทยหรือไม่ + isthai(text,check_all=False) + text คือ ข้อความหรือ list ตัวอักษร + check_all สำหรับส่งคืนค่า True หรือ False เช็คทุกตัวอักษร + + การส่งคืนค่า + {'thai':% อักษรภาษาไทย,'check_all':tuple โดยจะเป็น (ตัวอักษร,True หรือ False)} + """ + listext=list(text) + i=0 + num_isthai=0 + if check_all==True: + listthai=[] + while i= 3584 and cVal <= 3711): + num_isthai+=1 + if check_all==True: + listthai.append(True) + else: + if check_all==True: + listthai.append(False) + i+=1 + thai=(num_isthai/len(listext))*100 + if check_all==True: + dictthai=tuple(zip(listext,listthai)) + data= {'thai':thai,'check_all':dictthai} + else: + data= {'thai':thai} + return data +def syllable_tokenize(text1): + """ + syllable_tokenize(text) + เป็นคำสั่งสำหรับใช้ตัดพยางค์ในภาษาไทย + รับ str + ส่งออก list + """ + text1=word_tokenize(text1) + data=[] + trie = create_custom_dict_trie(custom_dict_source=get_data()) + if(len(text1)>0): + i=0 + while(i>> from pythainlp.tokenize import dict_word_tokenize,create_custom_dict_trie + >>> listword=['แมว',"ดี"] + >>> data_dict=create_custom_dict_trie(listword) + >>> dict_word_tokenize("แมวดีดีแมว",data_dict) + ['แมว', 'ดี', 'ดี', 'แมว'] + ''' + if engine=="newmm" or engine=="onecut": + from .newmm import mmcut as segment + elif engine=="mm" or engine=="multi_cut": + from .multi_cut import segment + elif engine=='longest-matching': + from .longest import segment + elif engine=='wordcutpy': + from .wordcutpy import segment + return segment(text, custom_dict_trie.keys()) + else: + raise Exception("error no have engine.") + return segment(text, custom_dict_trie) def sent_tokenize(text,engine='whitespace+newline'): ''' - sent_tokenize(text,engine='whitespace+newline') - ตัดประโยคเบื้องต้น โดยการแบ่งด้วยช่องว่าง + This function does not yet automatically recognize when a sentence actually ends. Rather it helps split text where white space and a new line is found. + + :param str text: the text to be tokenized + :param str engine: choose between 'whitespace' or 'whitespace+newline' + + :return: a list of text, split by whitespace or new line. ''' if engine=='whitespace': data=nltk.tokenize.WhitespaceTokenizer().tokenize(text) elif engine=='whitespace+newline': data=re.sub(r'\n+|\s+','|',text,re.U).split('|') return data -def wordpunct_tokenize(text): - ''' - wordpunct_tokenize(text) - It is nltk.tokenize.wordpunct_tokenize(text). - ''' - return nltk.tokenize.wordpunct_tokenize(text) -def WhitespaceTokenizer(text): - return nltk.tokenize.WhitespaceTokenizer().tokenize(text) -def isthai(text,check_all=False): - """ - สำหรับเช็คว่าเป็นตัวอักษรภาษาไทยหรือไม่ - isthai(text,check_all=False) - text คือ ข้อความหรือ list ตัวอักษร - check_all สำหรับส่งคืนค่า True หรือ False เช็คทุกตัวอักษร - การส่งคืนค่า - {'thai':% อักษรภาษาไทย,'check_all':tuple โดยจะเป็น (ตัวอักษร,True หรือ False)} +def subword_tokenize(text, engine='tcc'): + """ + :param str text: text to be tokenized + :param str engine: choosing 'tcc' uses the Thai Character Cluster rule to segment words into the smallest unique units. + :return: a list of tokenized strings. """ - listext=list(text) - i=0 - num_isthai=0 - if check_all==True: - listthai=[] - while i= 3584 and cVal <= 3711): - num_isthai+=1 - if check_all==True: - listthai.append(True) - else: - if check_all==True: - listthai.append(False) - i+=1 - thai=(num_isthai/len(listext))*100 - if check_all==True: - dictthai=tuple(zip(listext,listthai)) - data= {'thai':thai,'check_all':dictthai} - else: - data= {'thai':thai} - return data -def syllable_tokenize(text1): + if engine == 'tcc': + from .tcc import tcc + return tcc(text) + +def isthai(text,check_all=False): """ - syllable_tokenize(text) - เป็นคำสั่งสำหรับใช้ตัดพยางค์ในภาษาไทย - รับ str - ส่งออก list + :param str text: input string or list of strings + :param bool check_all: checks all character or not + + :return: A dictionary with the first value as proportional of text that is Thai, and the second value being a tuple of all characters, along with true or false. """ - text1=word_tokenize(text1) + listext=list(text) + i=0 + num_isthai=0 + if check_all==True: + listthai=[] + while i= 3584 and cVal <= 3711): + num_isthai+=1 + if check_all==True: + listthai.append(True) + else: + if check_all==True: + listthai.append(False) + i+=1 + thai=(num_isthai/len(listext))*100 + if check_all==True: + dictthai=tuple(zip(listext,listthai)) + data= {'thai':thai,'check_all':dictthai} + else: + data= {'thai':thai} + return data + +def syllable_tokenize(text): + """ + :param str text: input string to be tokenized + + :return: returns list of strings of syllables + """ + text1=word_tokenize(text) data=[] - if(len(text1)>0): + trie = create_custom_dict_trie(custom_dict_source=get_data()) + if len(text1)>1: i=0 - while(i= 3584 and cVal <= 3711): - return True - return False \ No newline at end of file diff --git a/pythainlp/tokenize/longest.py b/pythainlp/tokenize/longest.py index 841d50fe5..6406c1665 100644 --- a/pythainlp/tokenize/longest.py +++ b/pythainlp/tokenize/longest.py @@ -3,8 +3,8 @@ ''' Code from https://github.com/patorn/thai-sentiment/blob/78bf461dfdc8a3f0517712fac56dd921dc0f9dd6/thai_sentiment/tokenizer.py ''' +from pythainlp.tokenize import DEFAULT_DICT_TRIE import re -from pythainlp.tools import file_trie FRONT_DEP_CHAR = ['ะ', 'ั', 'า ', 'ำ', 'ิ', 'ี', 'ึ', 'ื', 'ุ', 'ู', 'ๅ', '็', '์', 'ํ'] REAR_DEP_CHAR = ['ั', 'ื', 'เ', 'แ', 'โ', 'ใ', 'ไ', 'ํ'] TONAL_CHAR = ['่', '้', '๊', '๋'] @@ -12,12 +12,8 @@ class Tokenizer(object): - def __init__(self,data=''): - if(data==''): - self._trie = file_trie(data="old") - else: - import marisa_trie - self._trie = marisa_trie.Trie(data) + def __init__(self, trie): + self._trie = trie @property def trie(self): @@ -112,9 +108,8 @@ def segment_text(self, text): def tokenize(self, text): tokens = self.segment_text(text) return tokens -def segment(s,data=''): +def segment(s, trie=None): """ตัดคำภาษาไทยด้วย Longest matching""" - return Tokenizer(data).tokenize(s) -if __name__ == "__main__": - s = 'สวัสดีชาวโลกเข้าใจกันไหมพวกคุณ โอเคกันไหม ยสยา ดีแล้วนะคุณเธอ' - print(segment(s)) \ No newline at end of file + if not trie: + trie = DEFAULT_DICT_TRIE + return Tokenizer(trie).tokenize(s) \ No newline at end of file diff --git a/pythainlp/tokenize/mm.py b/pythainlp/tokenize/mm.py deleted file mode 100644 index fa13d2d2b..000000000 --- a/pythainlp/tokenize/mm.py +++ /dev/null @@ -1,299 +0,0 @@ -# -*- coding: utf-8 -*- - -''' -based on algorithm from -http://www.aclweb.org/anthology/E14-4016 - -fork from https://github.com/narongdejsrn/pythaiwordcut -License: MIT -''' - -from __future__ import print_function -from six.moves import range,zip -import codecs -import re -from pythainlp.tools import file_trie -from pythainlp.corpus import stopwords # load stopwords -import marisa_trie -class wordcut(object): - """ - ตัดคำภาษาไทยด้วย Maximum Matching algorithm - """ - def __init__(self, removeRepeat=True, keyDictionary="", stopDictionary="", removeSpaces=True, minLength=1, stopNumber=False, removeNonCharacter=False, caseSensitive=True, ngram=(1,1), negation=False,data=""): - if data!="": - d = data # load dictionary - # load negation listdir - self.negationDict = [] - if negation: - self.negationDict = ['ไม่','แต่'] - self.stopword = False - self.stopdict = [] - if(stopDictionary is not ""): - self.stopword = True - with codecs.open(stopDictionary, 'r',encoding='utf8') as f: - for line in f: - self.stopdict.append(line) - else: - self.stopdict = stopwords.words('thai') - self.keyword = False - self.keydict = [] - if(keyDictionary is not ""): - self.keyword = True - with codecs.open(keyDictionary, 'r',encoding='utf8') as f: - for line in f.read().splitlines(): - self.keydict.append(line) - - if data=="": - self.trie = file_trie(data="old") - else: - self.trie = marisa_trie.Trie(d) - self.removeRepeat = removeRepeat - self.stopNumber = stopNumber - self.removeSpaces = removeSpaces - self.minLength = minLength - self.removeNonCharacter = removeNonCharacter - self.caseSensitive = caseSensitive - self.ngram = ngram - self.negation = negation - self.onNegation = False - - def determine(self, word): - if self.stopNumber and word.isdigit(): - return False - - if self.removeSpaces and word.isspace(): - return False - - if len(word) < self.minLength: - return False - - if self.removeNonCharacter: - match = re.search(u"[0-9A-Za-z\u0E00-\u0E7F]+", word,re.U) - if not match: - return False - - return True - - # Find maximum matching in Trie if match return id else return -1 - def search_trie(self, word): - # remove negation if see a space - if(word[0:1] == " "): - self.onNegation = False - - # check latin words - match = re.search(u"[A-Za-z\d]*", word,re.U) - if match.group(0): - if not self.caseSensitive: - return match.group(0).lower() - else: - return match.group(0) - - # check number - match = re.search(u"[\d]*", word,re.U) - if match.group(0): - return match.group(0) - - longest = 0 - max_data = None - - for x in range(20): - if word[0:x] in self.trie: - longest = len(word[0:x]) - max_data = word[0:x] - - if longest > 20: - for data in self.trie.keys(word[0:longest]): - if len(data) > longest and data in word[0:len(data)]: - longest = len(data) - max_data = data - - - if max_data: - try: - # Special check for case like ๆ - if word[len(max_data)] == 'ๆ': - return word[0:(len(max_data) + 1)] - else: - return max_data - except: - return max_data - else: - return -1 - - def transform(self, wordArray): - for dd in self.stopdict: - try: - if self.caseSensitive: - wordArray.remove(dd) - else: - wordArray.remove(dd.lower()) - except ValueError: - pass - - return wordArray - - def extract_keyword(self, wordArray): - """ - ใช้ในการหาคำสำคัญ - """ - result_array = [] - for dd in wordArray: - try: - if self.caseSensitive and dd in self.keydict: - result_array.append(dd) - else: - if dd.lower() in self.keydict: - result_array.append(dd) - except ValueError: - pass - - return result_array - # c = sentence which represent as char - # N = number of character - def find_segment(self, c): - i = 0 - N = len(c) - arr = [] - while(i < N): - j = self.search_trie(c[i:N]) - if(j == -1): - if(self.removeRepeat is False or c[i] != c[i - 1]): - arr.append(c[i]) - i = i + 1 - else: - i = i + 1 - else: - k = j - if self.negation: - if self.onNegation: - k = 'NOT_' + j - - if j in self.negationDict: - self.onNegation = True - - arr.append(k) - i = i + len(j) - return arr - - def find_ngrams(self, input_list, n): - return zip(*[input_list[i:] for i in range(n)]) - - def segment(self, c): - ''' - ตัดคำใช้ฟังก์ชัน segment - ''' - result = self.find_segment(c) - if self.stopword: - result = self.transform(result) - - result = [x for x in result if self.determine(x)] - - lastresult = [] - for x in range(self.ngram[0], self.ngram[1]+1): - for r in self.find_ngrams(result, x): - match = re.search(u"[A-Za-z\d]+", ''.join(r),re.U) - if not match: - lastresult.append(''.join(r)) - else: - if self.negation: - lastresult.append(''.join(r)) - else: - lastresult.append(' '.join(r)) - return lastresult -def mergelistlen(listdata,lennum): - ''' - แก้ Bug ที่เกิดจาก mm - ''' - i=0 - listlen=len(listdata) - while i(listlen-1) or i+1==listlen: - ''' - ถ้า i เกินความยาว list ให้ออกจากการลูป - ''' - break - elif re.search(r'[0-9]',listdata[i]): - ''' - ถ้าหาก listdata[i] เป็นตัวเลขให้ข้ามไป - ''' - pass - elif re.search(r'[ะา]',listdata[i]) and (len(listdata[i])==lennum and len(listdata[i+1])==lennum): - ''' - ถ้าหาก listdata[i] คือ ะ/า ซึ่งเปนสระที่ไว้ข้างหลังได้เท่านั้น และ listdata[i] กับ listdata[i+1] ยาวเท่า lennum - จึงนำ listdata[i] ไปรวมกับ listdata[i-1] แล้วลบ listdata[i] ออก - ''' - listdata[i-1]+=listdata[i] - del listdata[i] - i-=1 - elif re.search(r'[ก-ฮ]',listdata[i]) and re.search(r'[0-9]',listdata[i+1]): - ''' - กันปัญหา ก-ฮ ตัวต่อมาเป็น 0-9 มีความยาวเท่ากัน ให้ ก-ฮ ไปรวมกับตัวก่อนหน้า - ''' - listdata[i-1]+=listdata[i] - del listdata[i] - i-=1 - elif len(listdata[i])==lennum and len(listdata[i+1])==lennum: - ''' - ถ้าหาก list มีความยาวเท่ากันอยู่ติดกัน - ''' - #print(listdata,'99') - T=True - num=1 - while T==True: - if (i+num)>=listlen: - ii=i - num2=1 - TT=True - while TT==True: - if (i+num2)<=(listlen-1): - listdata[i]+=listdata[i+num2] - num2+=1 - elif (i+num2)>(listlen-1): - num2-=1 - TT=False - TT=True - while TT==True: - if (i+num2) != i: - del listdata[i+num2] - num2-=1 - else: - TT=False - T=False - elif len(listdata[i+(num-1)])!=len(listdata[i+num]): #and re.search(r'[0-9]',listdata[i+(num-1)])==False:# and isThai(listdata[i+(num-1)])==True: - ii=1+i - while ii<(i+num) and ii<(len(listdata)-1): - listdata[i]+=listdata[ii] - ii+=1 - ii=i+num-1 - while ii>i: - del listdata[ii] - ii-=1 - T=False - num+=1 - del T,ii - elif len(listdata[i])==lennum and len(listdata[i+1])!=lennum: - ''' - ในกรณีที่ list ความยาวที่กำหนด แต่ตัวต่อไปยาวไม่เท่า ให้ยุบรวมกัน - ''' - if re.search(r'[เแโใไ]',listdata[i]): - ''' - ถ้าหากเป็นสระต้นคำ ให้รวมกัน - ''' - listdata[i]+=listdata[i+1] - del listdata[i+1] - elif re.search(r'[ก-ฮ]',listdata[i]) or re.search(r'[ะา]',listdata[i]): - ''' - หากเป็นแค่พยัญชนะให้รวมกับตัวหลัง - ''' - listdata[i-1]+=listdata[i] - del listdata[i] - i-=1 - listlen=len(listdata) - i+=1 - return listdata -def segment(text,data=""): - ''' - ใช้ในการตัดตำ segment(str) คืนค่า list - ''' - pt = wordcut(stopNumber=False, removeNonCharacter=True, caseSensitive=False,removeRepeat=True,data=data) - return mergelistlen(pt.segment(text),1) \ No newline at end of file diff --git a/pythainlp/tokenize/multi_cut.py b/pythainlp/tokenize/multi_cut.py new file mode 100644 index 000000000..f8d1cd45f --- /dev/null +++ b/pythainlp/tokenize/multi_cut.py @@ -0,0 +1,121 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,division,unicode_literals,print_function +from builtins import * +''' +โปรแกรม multi-cut +ตัดคำภาษาไทยโดยใช้ Maximum Matching algorithm +เดติดโค้ดต้นฉบับ คุณ Korakot Chaovavanich +จาก https://www.facebook.com/groups/408004796247683/permalink/431283740586455/ +และ https://gist.github.com/korakot/fe26c65dc9eed467f4497f784a805716 +''' +import re +from marisa_trie import Trie +from collections import defaultdict +from pythainlp.corpus.thaiword import get_data +DEFAULT_DICT_TRIE = Trie(get_data()) +class LatticeString(str): + ''' String subclass เพื่อเก็บวิธีตัดหลายๆ วิธี + ''' + def __new__(cls, value, multi=None, in_dict=True): + return str.__new__(cls, value) + + def __init__(self, value, multi=None, in_dict=True): + self.unique = True + if multi: + self.multi = list(multi) + if len(self.multi) > 1: + self.unique = False + else: + self.multi = [value] + self.in_dict = in_dict # บอกว่าเป็นคำมีในดิกหรือเปล่า + +spat_eng = r'''(?x) +[-a-zA-Z]+| # english +\d[\d,\.]*| # number +[ \t]+| # space +\r?\n # newline +''' +pat_eng = re.compile(spat_eng) + +def multicut(text,trie=None): + ''' ส่งคืน LatticeString คืนมาเป็นก้อนๆ + ''' + if not trie: + trie = DEFAULT_DICT_TRIE + words_at = defaultdict(list) # main data structure + def serialize(p, p2): # helper function + for w in words_at[p]: + p_ = p + len(w) + if p_== p2: + yield w + elif p_ < p2: + for path in serialize(p_, p2): + yield w+'/'+path + + q = {0} + last_p = 0 # last position for yield + while min(q) < len(text): + p = min(q) + q -= {p} # q.pop, but for set + + for w in trie.prefixes(text[p:]): + words_at[p].append(w) + q.add(p+len(w)) + + if len(q)==1: + q0 = min(q) + yield LatticeString(text[last_p:q0], serialize(last_p, q0)) + last_p = q0 + + # กรณี len(q) == 0 คือ ไม่มีใน dict + if len(q)==0: + m = pat_eng.match(text[p:]) + if m: # อังกฤษ, เลข, ว่าง + i = p + m.span()[1] + else: # skip น้อยที่สุด ที่เป็นไปได้ + for i in range(p, len(text)): + ww = trie.prefixes(text[i:]) + m = pat_eng.match(text[i:]) + if ww or m: + break + else: + i = len(text) + w = text[p:i] + words_at[p].append(w) + yield LatticeString(w, in_dict=False) + last_p = i + q.add(i) + +def mmcut(text): + res = [] + for w in multicut(text): + mm = min(w.multi, key=lambda x: x.count('/')) + res.extend(mm.split('/')) + return res +def combine(ww): + if ww == []: + yield "" + else: + w = ww[0] + for tail in combine(ww[1:]): + if w.unique: + yield w+"|"+tail + else: + for m in w.multi: + yield m.replace("/","|")+"|"+tail +def segment(text, trie=None): + ''' + ใช้ในการหา list ที่สามารถตัดคำได้ทั้งหมด + ''' + ww = list(multicut(text, trie=trie)) + return ww +def find_all_segment(text, trie=None): + ''' + ใช้ในการหา list ที่สามารถตัดคำได้ทั้งหมด + ''' + ww = list(multicut(text, trie=trie)) + return list(combine(ww)) +if __name__ == "__main__": + text='ผมรักคุณนะครับโอเคบ่พวกเราเป็นคนไทยรักภาษาไทยภาษาบ้านเกิด' + print(mmcut(text)) + #print(listcut(text)) diff --git a/pythainlp/tokenize/newmm.py b/pythainlp/tokenize/newmm.py index fde0471b6..66d3790c9 100644 --- a/pythainlp/tokenize/newmm.py +++ b/pythainlp/tokenize/newmm.py @@ -9,9 +9,7 @@ import re from collections import defaultdict from heapq import heappush, heappop # for priority queue -from marisa_trie import Trie -from pythainlp.corpus.thaiword import get_data # ดึงข้อมูลรายการคำในภาษาไทย - +from pythainlp.tokenize import DEFAULT_DICT_TRIE # ช่วยตัดพวกภาษาอังกฤษ เป็นต้น pat_eng = re.compile(r'''(?x) @@ -48,8 +46,6 @@ [เ-ไ]ct """.replace('c', '[ก-ฮ]').replace('t', '[่-๋]?').split() -THAI_WORDS = Trie(get_data()) - def tcc(w): p = 0 @@ -82,11 +78,7 @@ def bfs_paths_graph(graph, start, goal): else: queue.append((next, path+[next])) -def onecut(text, data=['']): - if(data != ['']): - trie = Trie(data) - else: - trie = THAI_WORDS +def onecut(text, trie): graph = defaultdict(list) # main data structure allow_pos = tcc_pos(text) # ตำแหน่งที่ตัด ต้องตรงกับ tcc @@ -120,6 +112,7 @@ def onecut(text, data=['']): for i in range(p+1, len(text)): if i in allow_pos: # ใช้ tcc ด้วย ww = [w for w in trie.prefixes(text[i:]) if (i+len(w) in allow_pos)] + ww = [w for w in ww if not re.match('[ก-ฮ]{,2}$', w)] m = pat_eng.match(text[i:]) if ww or m: break @@ -134,5 +127,7 @@ def onecut(text, data=['']): # ช่วยให้ไม่ต้องพิมพ์ยาวๆ -def mmcut(text, data=['']): - return list(onecut(text, data=data)) +def mmcut(text, trie=None): + if not trie: + trie = DEFAULT_DICT_TRIE + return list(onecut(text, trie)) diff --git a/pythainlp/tokenize/pyicu.py b/pythainlp/tokenize/pyicu.py index 57934568c..3b8f5ece6 100644 --- a/pythainlp/tokenize/pyicu.py +++ b/pythainlp/tokenize/pyicu.py @@ -1,80 +1,27 @@ # -*- coding: utf-8 -*- from __future__ import absolute_import,print_function,unicode_literals import sys -from six.moves import zip +import re try: import icu except ImportError: - import pip - pip.main(['install','pyicu']) + from pythainlp.tools import install_package + install_package('pyicu') try: import icu except ImportError: sys.exit('Error ! using pip install pyicu') -def isEnglish(s): - ''' - เช็คว่าตัวอักษรเป็นภาษาอังกฤษหรือไม่ - ''' - try: - try: - s.encode('ascii') - except UnicodeEncodeError: - return False - else: - return True - except: - try: - s.decode('ascii') - except UnicodeDecodeError: - return False - else: - return True -def isThai(chr1): - ''' - เช็คตัวอักษรว่าใช่ภาษาไทยไหม - ''' - if isEnglish(chr1): - return False - else: - try: - cVal = ord(chr1) - if(cVal >= 3584 and cVal <= 3711): - return True - else: - return False - except TypeError: - return False -def segment(txt): - """รับค่า ''str'' คืนค่าออกมาเป็น ''list'' ที่ได้มาจากการตัดคำโดย ICU""" - bd = icu.BreakIterator.createWordInstance(icu.Locale("th")) # เริ่มต้นเรียกใช้งานคำสั่งตัดคำโดยเลือกเป็นภาษาไทยโดยใช้ icu - txt=txt.replace('\n','') - bd.setText(txt) # ทำการตัดคำ - breaks = list(bd) - result=[txt[x[0]:x[1]] for x in zip([0]+breaks, breaks)] - result1=[] - for data in result: - data1=list(data) - data2=[] - for txt1 in data1: - if isThai(txt1)==True: - if len(data2)==0: - data2.append(txt1) - else: - if isThai(data2[data1.index(txt1)-1])==True: - data2.append(txt1) - else: - data2.append(','+txt1) - else: - if len(data2)==0: - data2.append(txt1) - else: - if isThai(data2[data1.index(txt1)-1])==True: - data2.append(','+txt1) - else: - data2.append(txt1) - data1=''.join(data2) - result1+=data1.split(',') - return [x for x in result1 if x != ' '] +def gen_words(text): + bd = icu.BreakIterator.createWordInstance(icu.Locale("th")) + bd.setText(text) + p = bd.first() + for q in bd: + yield text[p:q] + p = q + +def segment(text): + text = re.sub("([^\u0E00-\u0E7F\n ]+)"," \\1 ",text) + return list(gen_words(text)) if __name__ == "__main__": print(segment('ทดสอบระบบตัดคำด้วยไอซียู')) print(segment('ผมชอบพูดไทยคำ English')) diff --git a/pythainlp/tokenize/pylexto.py b/pythainlp/tokenize/pylexto.py index d61207838..3334ab6e8 100644 --- a/pythainlp/tokenize/pylexto.py +++ b/pythainlp/tokenize/pylexto.py @@ -4,8 +4,8 @@ try: from pylexto import LexTo except ImportError: - import pip - pip.main(['install','https://github.com/wannaphongcom/pylexto/archive/master.zip']) + from pythainlp.tools import install_package + install_package('https://github.com/wannaphongcom/pylexto/archive/master.zip') try: from pylexto import LexTo except ImportError: diff --git a/pythainlp/tokenize/wordcutpy.py b/pythainlp/tokenize/wordcutpy.py index f90aaf1fb..5111709e3 100644 --- a/pythainlp/tokenize/wordcutpy.py +++ b/pythainlp/tokenize/wordcutpy.py @@ -11,16 +11,17 @@ ''' ในกรณีที่ยังไม่ติดตั้ง wordcutpy ในระบบ ''' - import pip - pip.main(['install','wordcutpy']) + from pythainlp.tools import install_package + install_package('wordcutpy') try: from wordcut import Wordcut except ImportError: sys.exit('Error ! using $ pip install wordcutpy') -def segment(text,data=""): - if data=="": + +def segment(text, data=None): + if not data: wordcut = Wordcut.bigthai() else: word_list = list(set(data)) wordcut = Wordcut(word_list) - return wordcut.tokenize(text) + return wordcut.tokenize(text) \ No newline at end of file diff --git a/pythainlp/tools/.ipynb_checkpoints/__init__-checkpoint.py b/pythainlp/tools/.ipynb_checkpoints/__init__-checkpoint.py new file mode 100644 index 000000000..83847e460 --- /dev/null +++ b/pythainlp/tools/.ipynb_checkpoints/__init__-checkpoint.py @@ -0,0 +1,82 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,unicode_literals +import os +import dill +from pythainlp.tokenize import tcc +import marisa_trie +def get_path_db(): + path = os.path.join(get_path_pythainlp_data(), "db.json") + if not os.path.exists(path): + from tinydb import TinyDB + db=TinyDB(path) + #db.insert({'name': 'hi', 'version': '0.1','file':''}) + return path +def get_path_data(filename): + return os.path.join(get_path_pythainlp_data(), filename) +def get_path_pythainlp_data(): + path= os.path.join(os.path.expanduser("~"), 'pythainlp-data') + if not os.path.exists(path): + os.makedirs(path) + return path +def file_trie(data): + ''' + ใช้สร้างไฟล์ข้อมูลสำหรับระบบที่ใช้ trie + ''' + path = get_path_pythainlp_data() + if not os.path.exists(path): + os.makedirs(path) + if data=="newmm": + path = os.path.join(path, 'pythainlp_trie-tcc1.data') + elif data=="old": + path = os.path.join(path, 'pythainlp_trie2.data') + else: + path = os.path.join(path, 'pythainlp_trie2.data') + if not os.path.exists(path): + #ถ้าไม่มีไฟล์ + if data=="newmm": + from pythainlp.corpus.thaiword import get_data # ข้อมูลเก่า + data2=get_data() + i=0 + while i', re.IGNORECASE) + self.re_rep = re.compile(r'(\S)(\1{3,})') + + def sub_br(self,x): return self.re_br.sub("\n", x) + + def tokenize(self,x): + return [t for t in word_tokenize(self.sub_br(x),engine=self.engine)] + + #replace aaaaaaaa + @staticmethod + def replace_rep(m): + TK_REP = 'tkrep' + c,cc = m.groups() + return f'{TK_REP}{len(cc)+1}{c}' + + def proc_text(self, s): + s = self.re_rep.sub(ThaiTokenizer.replace_rep, s) + s = re.sub(r'([/#])', r' \1 ', s) + #remvoe double space + s = re.sub(' {2,}', ' ', s) + return self.tokenize(s) + + @staticmethod + def proc_all(ss): + tok = ThaiTokenizer() + return [tok.proc_text(s) for s in ss] + + @staticmethod + def proc_all_mp(ss): + ncpus = num_cpus()//2 + with ProcessPoolExecutor(ncpus) as e: + return sum(e.map(ThaiTokenizer.proc_all, ss), []) + +#get tokenized texts +BOS = 'xbos' # beginning-of-sentence tag +def get_texts(df): + labels = df.iloc[:,0].values.astype(np.int64) + texts = BOS+df.iloc[:,1].astype(str).apply(lambda x: x.rstrip()) + tok = ThaiTokenizer().proc_all_mp(partition_by_cores(texts)) + return tok, list(labels) + +#get all tokenized texts +def get_all(df): + tok, labels = [], [] + for i, r in enumerate(df): + tok_, labels_ = get_texts(r) + tok += tok_; + labels += labels_ + return tok, labels + +#convert text dataframe to numericalized dataframes +def numericalizer(df, itos=None, max_vocab = 60000, min_freq = 2, pad_tok = '_pad_', unk_tok = '_unk_'): + tok, labels = get_all(df) + freq = Counter(p for o in tok for p in o) + if itos is None: + itos = [o for o,c in freq.most_common(max_vocab) if c>min_freq] + itos.insert(0, pad_tok) + itos.insert(0, unk_tok) + stoi = collections.defaultdict(lambda:0, {v:k for k,v in enumerate(itos)}) + lm = np.array([[stoi[o] for o in p] for p in tok]) + return(lm,tok,labels,itos,stoi,freq) + +#get document vectors from language model +def document_vector(ss, m, stoi,tok_engine='newmm'): + s = word_tokenize(ss) + t = LongTensor([stoi[i] for i in s]).view(-1,1).cuda() + t = Variable(t,volatile=False) + m.reset() + pred,*_ = m[0](t) + #get average of last lstm layer along bptt + res = to_np(torch.mean(pred[-1],0).view(-1)) + return(res) + +#merge pretrained embeddings with current embeddings +def merge_wgts(em_sz, wgts, itos_pre, itos_cls): + vocab_size = len(itos_cls) + enc_wgts = to_np(wgts['0.encoder.weight']) + #average weight of encoding + row_m = enc_wgts.mean(0) + stoi_pre = collections.defaultdict(lambda:-1, {v:k for k,v in enumerate(itos_pre)}) + #new embedding based on classification dataset + new_w = np.zeros((vocab_size, em_sz), dtype=np.float32) + for i,w in enumerate(itos_cls): + r = stoi_pre[w] + #use pretrianed embedding if present; else use the average + new_w[i] = enc_wgts[r] if r>=0 else row_m + wgts['0.encoder.weight'] = T(new_w) + wgts['0.encoder_with_dropout.embed.weight'] = T(np.copy(new_w)) + wgts['1.decoder.weight'] = T(np.copy(new_w)) + return(wgts) + +#feature extractor +class SaveFeatures(): + features=None + def __init__(self, m): self.hook = m.register_forward_hook(self.hook_fn) + def hook_fn(self, module, input, output): self.features = output + def remove(self): self.hook.remove() + +#download pretrained language model +def get_path(fname): + path = get_file(fname) + if path==None: + download(fname) + path = get_file(fname) + return(path) + +def load_pretrained_model(): + path = get_path(MODEL_NAME) + wgts = torch.load(path, map_location=lambda storage, loc: storage) + return(wgts) + +def load_pretrained_itos(): + path = get_path(ITOS_NAME) + itos = pickle.load(open(path,'rb')) + return(itos) + +def about(): + return ''' + thai2vec + State-of-the-Art Language Modeling, Text Feature Extraction and Text Classification in Thai Language. + Created as part of pyThaiNLP with ULMFit implementation from fast.ai + + Development : Charin Polpanumas + GitHub : https://github.com/cstorm125/thai2vec + ''' + + diff --git a/pythainlp/ulmfit/__init__ b/pythainlp/ulmfit/__init__ new file mode 100644 index 000000000..22e30a360 --- /dev/null +++ b/pythainlp/ulmfit/__init__ @@ -0,0 +1,3 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,unicode_literals +from .thai2vec import * \ No newline at end of file diff --git a/pythainlp/ulmfit/__init__.py b/pythainlp/ulmfit/__init__.py new file mode 100644 index 000000000..d61c3dfd8 --- /dev/null +++ b/pythainlp/ulmfit/__init__.py @@ -0,0 +1,2 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,unicode_literals \ No newline at end of file diff --git a/pythainlp/ulmfit/utils.py b/pythainlp/ulmfit/utils.py new file mode 100644 index 000000000..dcb19ba4b --- /dev/null +++ b/pythainlp/ulmfit/utils.py @@ -0,0 +1,258 @@ +# -*- coding: utf-8 -*- +''' +Code by https://github.com/cstorm125/thai2vec/tree/master/notebook +''' +from __future__ import absolute_import,unicode_literals +import os +import sys +import re +import torch + +#numpy and fastai +try: + import numpy as np + from fastai.text import * + import dill as pickle +except ImportError: + from pythainlp.tools import install_package + install_package('fastai') + install_package('numpy') + try: + import numpy as np + from fastai.text import * + import dill as pickle + except ImportError: + print("Error installing using 'pip install fastai numpy dill'") + sys.exit(0) + +#import torch +try: + import torch +except ImportError: + print('PyTorch required. See https://pytorch.org/.') + +from pythainlp.tokenize import word_tokenize +from pythainlp.corpus import get_file +from pythainlp.corpus import download +MODEL_NAME = 'thwiki_model2' +ITOS_NAME = 'itos' + +#paralellized thai tokenizer with some text cleaning +class ThaiTokenizer(): + def __init__(self, engine='newmm'): + """ + :parameters for tokenization engine: + * newmm - Maximum Matching algorithm + TCC + * icu - IBM ICU + * longest-matching - Longest matching + * mm - Maximum Matching algorithm + * pylexto - LexTo + * deepcut - Deep Neural Network + """ + self.engine = engine + self.re_br = re.compile(r'<\s*br\s*/?>', re.IGNORECASE) + self.re_rep = re.compile(r'(\S)(\1{3,})') + + def sub_br(self,text): + """ + :meth:`sub_br` replace `
` tags with `\n` + :param str text: text to process + :return: procssed text + """ + return self.re_br.sub("\n", text) + + def tokenize(self,text): + """ + :meth: tokenize text with selected engine + :param str text: text to tokenize + :return: tokenized text + """ + return [t for t in word_tokenize(self.sub_br(text),engine=self.engine)] + + @staticmethod + def replace_rep(text): + ''' + :meth:`replace_rep` replace 3 or above repetitive characters with `tkrep` + :param str text: text to process + :return: processed text where repetitions are replaced by `tkrep` followed by number of repetitions + **Example**:: + >>> from pythainlp.ulmfit.utils import ThaiTokenizer + >>> tt = ThaiTokenizer() + >>> tt.replace_rep('คือดียยยยยย') + คือดีtkrep6ย + ''' + TK_REP = 'tkrep' + c,cc = text.groups() + return f'{TK_REP}{len(cc)+1}{c}' + + def proc_text(self, text): + """ + :meth: `proc_text` procss and tokenize text removing repetitions, special characters, double spaces + :param str text: text to process + :return: processed and tokenized text + """ + s = self.re_rep.sub(ThaiTokenizer.replace_rep, text) + s = re.sub(r'([/#])', r' \1 ', s) + #remvoe double space + s = re.sub(' {2,}', ' ', s) + return self.tokenize(s) + + @staticmethod + def proc_all(ss): + """ + :meth: `proc_all` runs `proc_text` for multiple sentences + :param str text: text to process + :return: processed and tokenized text + """ + tok = ThaiTokenizer() + return [tok.proc_text(s) for s in ss] + + @staticmethod + def proc_all_mp(ss): + """ + :meth: `proc_all` runs `proc_text` for multiple sentences using multiple cpus + :param str text: text to process + :return: processed and tokenized text + """ + ncpus = num_cpus()//2 + with ProcessPoolExecutor(ncpus) as e: + return sum(e.map(ThaiTokenizer.proc_all, ss), []) + +#ulmfit helper functions +BOS = 'xbos' # beginning-of-sentence tag +def get_texts(df): + """ + :meth: `get_texts` get tuple of tokenized texts and labels + :param pandas.DataFrame df: `pandas.DataFrame` with `label` as first column and `text` as second column + :return: + * tok - lists of tokenized texts with beginning-of-sentence tag `xbos` as first element of each list + * labels - list of labels + """ + labels = df.iloc[:,0].values.astype(np.int64) + texts = BOS+df.iloc[:,1].astype(str).apply(lambda x: x.rstrip()) + tok = ThaiTokenizer().proc_all_mp(partition_by_cores(texts)) + return(tok, list(labels)) + +def get_all(df): + """ + :meth: `get_all` iterate `get_texts` for all the entire `pandas.DataFrame` + :param pandas.DataFrame df: `pandas.DataFrame` with `label` as first column and `text` as second column + :return: + * tok - lists of tokenized texts with beginning-of-sentence tag `xbos` as first element of each list + * labels - list of labels + """ + tok, labels = [], [] + for i, r in enumerate(df): + tok_, labels_ = get_texts(r) + tok += tok_; + labels += labels_ + return(tok, labels) + +def numericalizer(df, itos=None, max_vocab = 60000, min_freq = 2, pad_tok = '_pad_', unk_tok = '_unk_'): + """ + :meth: `numericalize` numericalize tokenized texts for: + * tokens with word frequency more than `min_freq` + * at maximum vocab size of `max_vocab` + * add unknown token `_unk_` and padding token `_pad_` in first and second position + * use integer-to-string list `itos` if avaiable e.g. ['_unk_', '_pad_','first_word','second_word',...] + :param pandas.DataFrame df: `pandas.DataFrame` with `label` as first column and `text` as second column + :param list itos: integer-to-string list + :param int max_vocab: maximum number of vocabulary (default 60000) + :param int min_freq: minimum word frequency to be included (default 2) + :param str pad_tok: padding token + :param str unk_token: unknown token + :return: + * lm - `numpy.array` of numericalized texts + * tok - lists of tokenized texts with beginning-of-sentence tag `xbos` as first element of each list + * labels - list of labels + * itos - integer-to-string list e.g. ['_unk_', '_pad_','first_word','second_word',...] + * stoi - string-to-integer dict e.g. {'_unk_':0, '_pad_':1,'first_word':2,'second_word':3,...} + * freq - `collections.Counter` for word frequency + """ + tok, labels = get_all(df) + freq = Counter(p for o in tok for p in o) + if itos is None: + itos = [o for o,c in freq.most_common(max_vocab) if c>min_freq] + itos.insert(0, pad_tok) + itos.insert(0, unk_tok) + stoi = collections.defaultdict(lambda:0, {v:k for k,v in enumerate(itos)}) + lm = np.array([[stoi[o] for o in p] for p in tok]) + return(lm,tok,labels,itos,stoi,freq) + +def merge_wgts(em_sz, wgts, itos_pre, itos_cls): + """ + :param pandas.DataFrame df: `pandas.DataFrame` with `label` as first column and `text` as second column + :param int em_sz: size of embedding vectors (pretrained model is at 300) + :param wgts: saved pyTorch weights of pretrained model + :param list itos_pre: integer-to-string list of pretrained model + :param list itos_cls: integer-to-string list of current dataset + :return: merged weights of the model for current dataset + """ + vocab_size = len(itos_cls) + enc_wgts = to_np(wgts['0.encoder.weight']) + #average weight of encoding + row_m = enc_wgts.mean(0) + stoi_pre = collections.defaultdict(lambda:-1, {v:k for k,v in enumerate(itos_pre)}) + #new embedding based on classification dataset + new_w = np.zeros((vocab_size, em_sz), dtype=np.float32) + for i,w in enumerate(itos_cls): + r = stoi_pre[w] + #use pretrianed embedding if present; else use the average + new_w[i] = enc_wgts[r] if r>=0 else row_m + wgts['0.encoder.weight'] = T(new_w) + wgts['0.encoder_with_dropout.embed.weight'] = T(np.copy(new_w)) + wgts['1.decoder.weight'] = T(np.copy(new_w)) + return(wgts) + +#feature extractor +def document_vector(ss, m, stoi,tok_engine='newmm'): + """ + :meth: `document_vector` get document vector using pretrained ULMFit model + :param str ss: sentence to extract embeddings + :param m: pyTorch model + :param dict stoi: string-to-integer dict e.g. {'_unk_':0, '_pad_':1,'first_word':2,'second_word':3,...} + :param str tok_engine: tokenization engine (recommend using `newmm` if you are using pretrained ULMFit model) + :return: `numpy.array` of document vector sized 300 + """ + s = word_tokenize(ss) + t = LongTensor([stoi[i] for i in s]).view(-1,1).cuda() + t = Variable(t,volatile=False) + m.reset() + pred,*_ = m[0](t) + #get average of last lstm layer along bptt + res = to_np(torch.mean(pred[-1],0).view(-1)) + return(res) + +class SaveFeatures(): + features=None + def __init__(self, m): self.hook = m.register_forward_hook(self.hook_fn) + def hook_fn(self, module, input, output): self.features = output + def remove(self): self.hook.remove() + +#Download pretrained models +def get_path(fname): + path = get_file(fname) + if path==None: + download(fname) + path = get_file(fname) + return(path) + +def load_pretrained_model(): + path = get_path(MODEL_NAME) + wgts = torch.load(path, map_location=lambda storage, loc: storage) + return(wgts) + +def load_pretrained_itos(): + path = get_path(ITOS_NAME) + itos = pickle.load(open(path,'rb')) + return(itos) + +def about(): + return ''' + thai2vec + State-of-the-Art Language Modeling, Text Feature Extraction and Text Classification in Thai Language. + Created as part of pyThaiNLP with ULMFit implementation from fast.ai + + Development : Charin Polpanumas + GitHub : https://github.com/cstorm125/thai2vec + ''' diff --git a/pythainlp/util/__init__.py b/pythainlp/util/__init__.py index 45d9fc5da..332ef853d 100644 --- a/pythainlp/util/__init__.py +++ b/pythainlp/util/__init__.py @@ -94,4 +94,61 @@ def deletetone(data): search=re.findall(u'\w'+'์',data, re.U) for i in search: data=re.sub(i,'',data,flags=re.U) - return data \ No newline at end of file + return data +# Notebook : https://colab.research.google.com/drive/148WNIeclf0kOU6QxKd6pcfwpSs8l-VKD#scrollTo=EuVDd0nNuI8Q +# Cr. Korakot Chaovavanich +thaiword_nums = set('ศูนย์ หนึ่ง เอ็ด สอง ยี่ สาม สี่ ห้า หก เจ็ด แปด เก้า'.split()) +thaiword_units = set('สิบ ร้อย พัน หมื่น แสน ล้าน'.split()) +thaiword_nums_units = thaiword_nums | thaiword_units +thai_int_map = { + 'ศูนย์': 0, + 'หนึ่ง': 1, + 'เอ็ด': 1, + 'สอง': 2, + 'ยี่': 2, + 'สาม': 3, + 'สี่': 4, + 'ห้า': 5, + 'หก': 6, + 'เจ็ด': 7, + 'แปด': 8, + 'เก้า': 9, + 'สิบ': 10, + 'ร้อย': 100, + 'พัน': 1000, + 'หมื่น': 10000, + 'แสน': 100000, + 'ล้าน': 1000000, +} +nu_pat = re.compile('(.+)?(สิบ|ร้อย|พัน|หมื่น|แสน|ล้าน)(.+)?') # หกสิบ, ร้อยเอ็ด +# assuming that the units are separated already +def listtext_num2num_(tokens): + if len(tokens)==0: + return 0 + if len(tokens)==1: + return thai_int_map[tokens[0]] + if len(tokens)==2: + a, b = tokens + if b in thaiword_units: + return thai_int_map[a]*thai_int_map[b] + else: + return thai_int_map[a]+thai_int_map[b] + # longer case we use recursive + a, b = tokens[:2] + if a in thaiword_units and b != 'ล้าน': # ร้อย แปด + return thai_int_map[a] + listtext_num2num_(tokens[1:]) + # most common case, a isa num, b isa unit + if b in thaiword_units: + return thai_int_map[a]*thai_int_map[b] + listtext_num2num_(tokens[2:]) +def listtext_num2num(tokens): + res = [] + for tok in tokens: + if tok in thaiword_nums_units: + res.append(tok) + else: + m = nu_pat.fullmatch(tok) + if m: + res.extend([t for t in m.groups() if t]) # ตัด None ทิ้ง + else: + pass # should not be here + return listtext_num2num_(res) \ No newline at end of file diff --git a/pythainlp/word_vector/.ipynb_checkpoints/__init__-checkpoint b/pythainlp/word_vector/.ipynb_checkpoints/__init__-checkpoint new file mode 100644 index 000000000..22e30a360 --- /dev/null +++ b/pythainlp/word_vector/.ipynb_checkpoints/__init__-checkpoint @@ -0,0 +1,3 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,unicode_literals +from .thai2vec import * \ No newline at end of file diff --git a/pythainlp/word_vector/.ipynb_checkpoints/__init__-checkpoint.py b/pythainlp/word_vector/.ipynb_checkpoints/__init__-checkpoint.py new file mode 100644 index 000000000..d61c3dfd8 --- /dev/null +++ b/pythainlp/word_vector/.ipynb_checkpoints/__init__-checkpoint.py @@ -0,0 +1,2 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import,unicode_literals \ No newline at end of file diff --git a/pythainlp/word_vector/.ipynb_checkpoints/thai2vec-checkpoint.py b/pythainlp/word_vector/.ipynb_checkpoints/thai2vec-checkpoint.py new file mode 100644 index 000000000..264228d70 --- /dev/null +++ b/pythainlp/word_vector/.ipynb_checkpoints/thai2vec-checkpoint.py @@ -0,0 +1,64 @@ +# -*- coding: utf-8 -*- +''' +Code by https://github.com/cstorm125/thai2vec/blob/master/notebooks/examples.ipynb +''' +from __future__ import absolute_import,unicode_literals +import six +import sys +if six.PY2: + print("Thai sentiment in pythainlp. Not support python 2.7") + sys.exit(0) +try: + from gensim.models import KeyedVectors + import numpy as np +except ImportError: + import pip + pip.main(['install','gensim','numpy']) + try: + from gensim.models import KeyedVectors + import numpy as np + except ImportError: + print("Error ! using 'pip install gensim numpy'") + sys.exit(0) +from pythainlp.tokenize import word_tokenize +from pythainlp.corpus import get_file +from pythainlp.corpus import download as download_data +import os + +def download(): + path = get_file('thai2vec02') + if path==None: + download_data('thai2vec02') + path = get_file('thai2vec02') + return path +def get_model(): + return KeyedVectors.load_word2vec_format(download(),binary=False) +def most_similar_cosmul(positive,negative): + ''' + การใช้งาน + input list + ''' + return get_model().most_similar_cosmul(positive=positive, negative=negative) +def doesnt_match(listdata): + return get_model().doesnt_match(listdata) +def similarity(word1,word2): + return get_model().similarity(word1,word2) +def sentence_vectorizer(ss,dim=300,use_mean=False): + s = word_tokenize(ss) + vec = np.zeros((1,dim)) + for word in s: + if word in get_model().wv.index2word: + vec+= get_model().wv.word_vec(word) + else: pass + if use_mean: vec /= len(s) + return(vec) + +def about(): + return ''' + thai2vec + State-of-the-Art Language Modeling, Text Feature Extraction and Text Classification in Thai Language. + Created as part of pyThaiNLP with ULMFit implementation from fast.ai + + Development : Charin Polpanumas + GitHub : https://github.com/cstorm125/thai2vec + ''' diff --git a/pythainlp/word_vector/thai2vec.py b/pythainlp/word_vector/thai2vec.py index d59298313..103e80b12 100644 --- a/pythainlp/word_vector/thai2vec.py +++ b/pythainlp/word_vector/thai2vec.py @@ -12,8 +12,9 @@ from gensim.models import KeyedVectors import numpy as np except ImportError: - import pip - pip.main(['install','gensim','numpy']) + from pythainlp.tools import install_package + install_package('gensim') + install_package('numpy') try: from gensim.models import KeyedVectors import numpy as np @@ -21,31 +22,35 @@ print("Error ! using 'pip install gensim numpy'") sys.exit(0) from pythainlp.tokenize import word_tokenize +from pythainlp.corpus import get_file +from pythainlp.corpus import download as download_data import os def download(): - path = os.path.join(os.path.expanduser("~"), 'pythainlp-data') - if not os.path.exists(path): - os.makedirs(path) - path = os.path.join(path, 'thai2vec.vec') - if not os.path.exists(path): - print("Download models...") - from urllib import request - request.urlretrieve("https://www.dropbox.com/s/upnbmiebkfma7oy/thai2vec.vec?dl=1",path) - print("OK.") + path = get_file('thai2vec02') + if path==None: + download_data('thai2vec02') + path = get_file('thai2vec02') return path def get_model(): - return KeyedVectors.load_word2vec_format(download(),binary=False) + ''' + :return: Downloads the `gensim` model.''' + return KeyedVectors.load_word2vec_format(download(),binary=False) def most_similar_cosmul(positive,negative): - ''' + ''' การใช้งาน input list ''' - return get_model().most_similar_cosmul(positive=positive, negative=negative) + return get_model().most_similar_cosmul(positive=positive, negative=negative) def doesnt_match(listdata): - return get_model().doesnt_match(listdata) + return get_model().doesnt_match(listdata) def similarity(word1,word2): - return get_model().similarity(word1,word2) + ''' + :param str word1: first word + :param str word2: second word + :return: the cosine similarity between the two word vectors + ''' + return get_model().similarity(word1,word2) def sentence_vectorizer(ss,dim=300,use_mean=False): s = word_tokenize(ss) vec = np.zeros((1,dim)) @@ -55,10 +60,12 @@ def sentence_vectorizer(ss,dim=300,use_mean=False): else: pass if use_mean: vec /= len(s) return(vec) + def about(): return ''' thai2vec - Language Modeling, Word2Vec and Text Classification in Thai Language. Created as part of pyThaiNLP. + State-of-the-Art Language Modeling, Text Feature Extraction and Text Classification in Thai Language. + Created as part of pyThaiNLP with ULMFit implementation from fast.ai Development : Charin Polpanumas GitHub : https://github.com/cstorm125/thai2vec diff --git a/requirements-travis.txt b/requirements-travis.txt index 7596198d6..e373efe83 100644 --- a/requirements-travis.txt +++ b/requirements-travis.txt @@ -6,4 +6,7 @@ marisa_trie langdetect requests dill -pytz \ No newline at end of file +pytz +conllu +tinydb +tqdm diff --git a/requirements.txt b/requirements.txt index a94a74321..6a15aac4f 100644 --- a/requirements.txt +++ b/requirements.txt @@ -4,4 +4,7 @@ six marisa_trie requests dill -pytz \ No newline at end of file +pytz +conllu +tinydb +tqdm diff --git a/setup.cfg b/setup.cfg index 7bcce8927..42784cb6f 100644 --- a/setup.cfg +++ b/setup.cfg @@ -1,5 +1,5 @@ [bumpversion] -current_version = 1.6 +current_version = 1.7 commit = True tag = True @@ -13,6 +13,3 @@ replace = version='{new_version}' [bumpversion:file:pythainlp/__init__.py] search = __version__ = '{current_version}' replace = __version__ = '{new_version}' - -[bdist_wheel] -universal=1 diff --git a/setup.py b/setup.py index d4c23a343..a4d642a1d 100644 --- a/setup.py +++ b/setup.py @@ -1,33 +1,24 @@ # -*- coding: utf-8 -*- from setuptools import setup,find_packages import codecs -with codecs.open('README.rst','r',encoding='utf-8') as readme_file: +with codecs.open('README-pythainlp.md','r',encoding='utf-8') as readme_file: readme = readme_file.read() readme_file.close() -requirements = [ - 'nltk>=3.2.2', - 'future>=0.16.0', - 'six', - 'marisa_trie<=0.7.4', - 'requests', - 'dill', - 'pytz' -] -test_requirements = [ - # TODO: put package test requirements here -] +with codecs.open('requirements.txt','r',encoding='utf-8') as f: + requirements = f.read().splitlines() setup( name='pythainlp', - version='1.6.0.7', + version='1.7.0', description="Thai natural language processing in Python package.", long_description=readme, + long_description_content_type='text/markdown', author='PyThaiNLP', author_email='wannaphong@kkumail.com', url='https://github.com/PyThaiNLP/pythainlp', packages=find_packages(), test_suite='pythainlp.test', - package_data={'pythainlp.corpus':['stopwords-th.txt','thaipos.json','thaiword.txt','corpus_license.md','tha-wn.db','new-thaidict.txt','negation.txt','provinces.csv'],'pythainlp.sentiment':['vocabulary.data','sentiment.data']}, + package_data={'pythainlp.corpus':['stopwords-th.txt','thaipos.json','thaiword.txt','corpus_license.md','tha-wn.db','new-thaidict.txt','negation.txt','provinces.csv','pt_tagger_1.dill','ud_thai-pud_pt_tagger.dill','ud_thai-pud_unigram_tagger.dill','unigram_tagger.dill'],'pythainlp.sentiment':['vocabulary.data','sentiment.data']}, include_package_data=True, install_requires=requirements, license='Apache Software License 2.0', @@ -35,9 +26,9 @@ keywords='pythainlp', classifiers=[ 'Development Status :: 5 - Production/Stable', + 'Programming Language :: Python :: 3', 'Intended Audience :: Developers', 'License :: OSI Approved :: Apache Software License', 'Natural Language :: Thai', - 'Topic :: Text Processing :: Linguistic', - 'Programming Language :: Python :: Implementation'] + 'Topic :: Text Processing :: Linguistic'] )