desc
stringlengths 3
26.7k
| decl
stringlengths 11
7.89k
| bodies
stringlengths 8
553k
|
---|---|---|
'__init__: This has to be present'
| def __init__(self):
| self.a = 1
def XXX22():
'XXX22: This has to be present'
pass
|
'XXX22: This has to be present'
| def XXX11():
| pass
|
'Remove recursive references to allow garbage
collector to collect this object.'
| def cleanup(self):
| for dict in (self.rule2func, self.rules, self.rule2name):
for i in dict.keys():
dict[i] = None
for i in dir(self):
setattr(self, i, None)
|
'Disassemble a code object, returning a list of \'Token\'.
The main part of this procedure is modelled after
dis.disassemble().'
| def disassemble(self, co, classname=None, deob=0):
| rv = []
customize = {}
Token = self.Token
self.code = array('B', co.co_code)
linestarts = list(dis.findlinestarts(co))
varnames = list(co.co_varnames)
if deob:
linestarts = self.deobfuscate(co, linestarts, varnames)
code = self.code
n = len(code)
self.prev = [0]
for i in self.op_range(0, n):
op = code[i]
self.prev.append(i)
if (op >= HAVE_ARGUMENT):
self.prev.append(i)
self.prev.append(i)
self.lines = []
linetuple = namedtuple('linetuple', ['l_no', 'next'])
j = 0
linestartoffsets = {a for (a, _) in linestarts}
(prev_start_byte, prev_line_no) = linestarts[0]
for (start_byte, line_no) in linestarts[1:]:
while (j < start_byte):
self.lines.append(linetuple(prev_line_no, start_byte))
j += 1
last_op = code[self.prev[start_byte]]
(prev_start_byte, prev_line_no) = (start_byte, line_no)
while (j < n):
self.lines.append(linetuple(prev_line_no, n))
j += 1
if classname:
classname = (('_' + classname.lstrip('_')) + '__')
def unmangle(name):
if (name.startswith(classname) and (name[(-2):] != '__')):
return name[(len(classname) - 2):]
return name
free = [unmangle(name) for name in (co.co_cellvars + co.co_freevars)]
names = [unmangle(name) for name in co.co_names]
varnames = [unmangle(name) for name in varnames]
else:
free = (co.co_cellvars + co.co_freevars)
names = co.co_names
self.load_asserts = set()
for i in self.op_range(0, n):
if ((code[i] == PJIT) and (code[(i + 3)] == LOAD_GLOBAL)):
if (names[(code[(i + 4)] + (256 * code[(i + 5)]))] == 'AssertionError'):
self.load_asserts.add((i + 3))
cf = self.find_jump_targets(code)
last_stmt = self.next_stmt[0]
i = self.next_stmt[last_stmt]
replace = {}
while (i < (n - 1)):
if (self.lines[last_stmt].next > i):
if (code[last_stmt] == PRINT_ITEM):
if (code[i] == PRINT_ITEM):
replace[i] = 'PRINT_ITEM_CONT'
elif (code[i] == PRINT_NEWLINE):
replace[i] = 'PRINT_NEWLINE_CONT'
last_stmt = i
i = self.next_stmt[i]
imports = self.all_instr(0, n, (IMPORT_NAME, IMPORT_FROM, IMPORT_STAR))
if (len(imports) > 1):
last_import = imports[0]
for i in imports[1:]:
if (self.lines[last_import].next > i):
if (code[last_import] == IMPORT_NAME == code[i]):
replace[i] = 'IMPORT_NAME_CONT'
last_import = i
extended_arg = 0
for offset in self.op_range(0, n):
if (offset in cf):
k = 0
for j in cf[offset]:
rv.append(Token('COME_FROM', None, repr(j), offset=('%s_%d' % (offset, k))))
k += 1
op = code[offset]
opname = dis.opname[op]
oparg = None
pattr = None
if (op >= HAVE_ARGUMENT):
oparg = ((code[(offset + 1)] + (code[(offset + 2)] * 256)) + extended_arg)
extended_arg = 0
if (op == dis.EXTENDED_ARG):
extended_arg = (oparg * 65536L)
continue
if (op in dis.hasconst):
const = co.co_consts[oparg]
if (type(const) == types.CodeType):
oparg = const
if (const.co_name == '<lambda>'):
assert (opname == 'LOAD_CONST')
opname = 'LOAD_LAMBDA'
elif (const.co_name == '<genexpr>'):
opname = 'LOAD_GENEXPR'
elif (const.co_name == '<dictcomp>'):
opname = 'LOAD_DICTCOMP'
elif (const.co_name == '<setcomp>'):
opname = 'LOAD_SETCOMP'
pattr = (('<code_object ' + const.co_name) + '>')
else:
pattr = const
elif (op in dis.hasname):
pattr = names[oparg]
elif (op in dis.hasjrel):
pattr = repr(((offset + 3) + oparg))
elif (op in dis.hasjabs):
pattr = repr(oparg)
elif (op in dis.haslocal):
pattr = varnames[oparg]
elif (op in dis.hascompare):
pattr = dis.cmp_op[oparg]
elif (op in dis.hasfree):
pattr = free[oparg]
if (op in (BUILD_LIST, BUILD_TUPLE, BUILD_SET, BUILD_SLICE, UNPACK_SEQUENCE, MAKE_FUNCTION, CALL_FUNCTION, MAKE_CLOSURE, CALL_FUNCTION_VAR, CALL_FUNCTION_KW, CALL_FUNCTION_VAR_KW, DUP_TOPX, RAISE_VARARGS)):
if ((op == BUILD_TUPLE) and (code[self.prev[offset]] == LOAD_CLOSURE)):
continue
else:
opname = ('%s_%d' % (opname, oparg))
if (op != BUILD_SLICE):
customize[opname] = oparg
elif (op == JA):
target = self.get_target(offset)
if (target < offset):
if ((offset in self.stmts) and (code[(offset + 3)] not in (END_FINALLY, POP_BLOCK)) and (offset not in self.not_continue)):
opname = 'CONTINUE'
else:
opname = 'JUMP_BACK'
elif (op == LOAD_GLOBAL):
if (offset in self.load_asserts):
opname = 'LOAD_ASSERT'
elif (op == RETURN_VALUE):
if (offset in self.return_end_ifs):
opname = 'RETURN_END_IF'
if (offset not in replace):
rv.append(Token(opname, oparg, pattr, offset, linestart=(offset in linestartoffsets)))
else:
rv.append(Token(replace[offset], oparg, pattr, offset, linestart=(offset in linestartoffsets)))
if self.showasm:
out = self.out
for t in rv:
print >>out, t
print >>out
return (rv, customize)
|
'Find the first <instr> in the block from start to end.
<instr> is any python bytecode instruction or a list of opcodes
If <instr> is an opcode with a target (like a jump), a target
destination can be specified which must match precisely if exact
is True, or if exact is False, the instruction which has a target
closest to <target> will be returned.
Return index to it or None if not found.'
| def first_instr(self, start, end, instr, target=None, exact=True):
| code = self.code
assert ((start >= 0) and (end <= len(code)))
try:
(None in instr)
except:
instr = [instr]
pos = None
distance = len(code)
for i in self.op_range(start, end):
op = code[i]
if (op in instr):
if (target is None):
return i
dest = self.get_target(i, op)
if (dest == target):
return i
elif (not exact):
_distance = abs((target - dest))
if (_distance < distance):
distance = _distance
pos = i
return pos
|
'Find the last <instr> in the block from start to end.
<instr> is any python bytecode instruction or a list of opcodes
If <instr> is an opcode with a target (like a jump), a target
destination can be specified which must match precisely if exact
is True, or if exact is False, the instruction which has a target
closest to <target> will be returned.
Return index to it or None if not found.'
| def last_instr(self, start, end, instr, target=None, exact=True):
| code = self.code
if (not ((start >= 0) and (end <= len(code)))):
return None
try:
(None in instr)
except:
instr = [instr]
pos = None
distance = len(code)
for i in self.op_range(start, end):
op = code[i]
if (op in instr):
if (target is None):
pos = i
else:
dest = self.get_target(i, op)
if (dest == target):
distance = 0
pos = i
elif (not exact):
_distance = abs((target - dest))
if (_distance <= distance):
distance = _distance
pos = i
return pos
|
'Find all <instr> in the block from start to end.
<instr> is any python bytecode instruction or a list of opcodes
If <instr> is an opcode with a target (like a jump), a target
destination can be specified which must match precisely.
Return a list with indexes to them or [] if none found.'
| def all_instr(self, start, end, instr, target=None, include_beyond_target=False):
| code = self.code
assert ((start >= 0) and (end <= len(code)))
try:
(None in instr)
except:
instr = [instr]
result = []
for i in self.op_range(start, end):
op = code[i]
if (op in instr):
if (target is None):
result.append(i)
else:
t = self.get_target(i, op)
if (include_beyond_target and (t >= target)):
result.append(i)
elif (t == target):
result.append(i)
return result
|
'Find all <instr> in the block from start to end.
<instr> is any python bytecode instruction or a list of opcodes
If <instr> is an opcode with a target (like a jump), a target
destination can be specified which must match precisely.
Return a list with indexes to them or [] if none found.'
| def rem_or(self, start, end, instr, target=None, include_beyond_target=False):
| code = self.code
assert ((start >= 0) and (end <= len(code)))
try:
(None in instr)
except:
instr = [instr]
result = []
for i in self.op_range(start, end):
op = code[i]
if (op in instr):
if (target is None):
result.append(i)
else:
t = self.get_target(i, op)
if (include_beyond_target and (t >= target)):
result.append(i)
elif (t == target):
result.append(i)
pjits = self.all_instr(start, end, PJIT)
filtered = []
for pjit in pjits:
tgt = (self.get_target(pjit) - 3)
for i in result:
if ((i <= pjit) or (i >= tgt)):
filtered.append(i)
result = filtered
filtered = []
return result
|
'Return the next jump that was generated by an except SomeException:
construct in a try...except...else clause or None if not found.'
| def next_except_jump(self, start):
| if (self.code[start] == DUP_TOP):
except_match = self.first_instr(start, len(self.code), POP_JUMP_IF_FALSE)
if except_match:
jmp = self.prev[self.get_target(except_match)]
self.ignore_if.add(except_match)
self.not_continue.add(jmp)
return jmp
count_END_FINALLY = 0
count_SETUP_ = 0
for i in self.op_range(start, len(self.code)):
op = self.code[i]
if (op == END_FINALLY):
if (count_END_FINALLY == count_SETUP_):
assert (self.code[self.prev[i]] in (JA, JF, RETURN_VALUE))
self.not_continue.add(self.prev[i])
return self.prev[i]
count_END_FINALLY += 1
elif (op in (SETUP_EXCEPT, SETUP_WITH, SETUP_FINALLY)):
count_SETUP_ += 1
|
'Restrict pos to parent boundaries.'
| def restrict_to_parent(self, target, parent):
| if (not (parent['start'] < target < parent['end'])):
target = parent['end']
return target
|
'Detect structures and their boundaries to fix optimizied jumps
in python2.3+'
| def detect_structure(self, pos, op=None):
| code = self.code
if (op is None):
op = code[pos]
parent = self.structs[0]
start = parent['start']
end = parent['end']
for s in self.structs:
_start = s['start']
_end = s['end']
if ((_start <= pos < _end) and ((_start >= start) and (_end <= end))):
start = _start
end = _end
parent = s
origStructCount = len(self.structs)
if (op == SETUP_LOOP):
start = (pos + 3)
target = self.get_target(pos, op)
end = self.restrict_to_parent(target, parent)
if (target != end):
self.fixed_jumps[pos] = end
(line_no, next_line_byte) = self.lines[pos]
jump_back = self.last_instr(start, end, JA, next_line_byte, False)
if (jump_back and (jump_back != self.prev[end]) and (code[(jump_back + 3)] in (JA, JF))):
if ((code[self.prev[end]] == RETURN_VALUE) or ((code[self.prev[end]] == POP_BLOCK) and (code[self.prev[self.prev[end]]] == RETURN_VALUE))):
jump_back = None
if (not jump_back):
jump_back = (self.last_instr(start, end, RETURN_VALUE) + 1)
if (not jump_back):
return
if (code[self.prev[next_line_byte]] not in (PJIF, PJIT)):
loop_type = 'for'
else:
loop_type = 'while'
self.ignore_if.add(self.prev[next_line_byte])
target = next_line_byte
end = (jump_back + 3)
else:
if (self.get_target(jump_back) >= next_line_byte):
jump_back = self.last_instr(start, end, JA, start, False)
if ((end > (jump_back + 4)) and (code[end] in (JF, JA))):
if (code[(jump_back + 4)] in (JA, JF)):
if (self.get_target((jump_back + 4)) == self.get_target(end)):
self.fixed_jumps[pos] = (jump_back + 4)
end = (jump_back + 4)
elif (target < pos):
self.fixed_jumps[pos] = (jump_back + 4)
end = (jump_back + 4)
target = self.get_target(jump_back, JA)
if (code[target] in (FOR_ITER, GET_ITER)):
loop_type = 'for'
else:
loop_type = 'while'
test = self.prev[next_line_byte]
if (test == pos):
loop_type = 'while 1'
else:
self.ignore_if.add(test)
test_target = self.get_target(test)
if (test_target > (jump_back + 3)):
jump_back = test_target
self.not_continue.add(jump_back)
self.loops.append(target)
self.structs.append({'type': (loop_type + '-loop'), 'start': target, 'end': jump_back})
if ((jump_back + 3) != end):
self.structs.append({'type': (loop_type + '-else'), 'start': (jump_back + 3), 'end': end})
elif (op == SETUP_EXCEPT):
start = (pos + 3)
target = self.get_target(pos, op)
end = self.restrict_to_parent(target, parent)
if (target != end):
self.fixed_jumps[pos] = end
self.structs.append({'type': 'try', 'start': start, 'end': (end - 4)})
end_else = start_else = self.get_target(self.prev[end])
i = end
while (self.code[i] != END_FINALLY):
jmp = self.next_except_jump(i)
if (self.code[jmp] == RETURN_VALUE):
self.structs.append({'type': 'except', 'start': i, 'end': (jmp + 1)})
i = (jmp + 1)
else:
if (self.get_target(jmp) != start_else):
end_else = self.get_target(jmp)
if (self.code[jmp] == JF):
self.fixed_jumps[jmp] = (-1)
self.structs.append({'type': 'except', 'start': i, 'end': jmp})
i = (jmp + 3)
if (end_else != start_else):
r_end_else = self.restrict_to_parent(end_else, parent)
self.structs.append({'type': 'try-else', 'start': (i + 1), 'end': r_end_else})
self.fixed_jumps[i] = r_end_else
else:
self.fixed_jumps[i] = (i + 1)
elif (op in (PJIF, PJIT)):
start = (pos + 3)
target = self.get_target(pos, op)
rtarget = self.restrict_to_parent(target, parent)
pre = self.prev
if ((target != rtarget) and (parent['type'] == 'and/or')):
self.fixed_jumps[pos] = rtarget
return
if ((code[pre[target]] in (JUMP_IF_FALSE_OR_POP, JUMP_IF_TRUE_OR_POP, PJIF, PJIT)) and (target > pos)):
self.fixed_jumps[pos] = pre[target]
self.structs.append({'type': 'and/or', 'start': start, 'end': pre[target]})
return
if (op == PJIF):
match = self.rem_or(start, self.next_stmt[pos], PJIF, target)
match = self.remove_mid_line_ifs(match)
if match:
if ((code[pre[rtarget]] in (JF, JA)) and (pre[rtarget] not in self.stmts) and (self.restrict_to_parent(self.get_target(pre[rtarget]), parent) == rtarget)):
if ((code[pre[pre[rtarget]]] == JA) and self.remove_mid_line_ifs([pos]) and (target == self.get_target(pre[pre[rtarget]])) and ((pre[pre[rtarget]] not in self.stmts) or (self.get_target(pre[pre[rtarget]]) > pre[pre[rtarget]])) and (1 == len(self.remove_mid_line_ifs(self.rem_or(start, pre[pre[rtarget]], (PJIF, PJIT), target))))):
pass
elif ((code[pre[pre[rtarget]]] == RETURN_VALUE) and self.remove_mid_line_ifs([pos]) and (1 == len((set(self.remove_mid_line_ifs(self.rem_or(start, pre[pre[rtarget]], (PJIF, PJIT), target))) | set(self.remove_mid_line_ifs(self.rem_or(start, pre[pre[rtarget]], (PJIF, PJIT, JA), pre[rtarget], True))))))):
pass
else:
fix = None
jump_ifs = self.all_instr(start, self.next_stmt[pos], PJIF)
last_jump_good = True
for j in jump_ifs:
if (target == self.get_target(j)):
if ((self.lines[j].next == (j + 3)) and last_jump_good):
fix = j
break
else:
last_jump_good = False
self.fixed_jumps[pos] = (fix or match[(-1)])
return
else:
self.fixed_jumps[pos] = match[(-1)]
return
else:
if ((pos + 3) in self.load_asserts):
if (code[pre[rtarget]] == RAISE_VARARGS):
return
self.load_asserts.remove((pos + 3))
next = self.next_stmt[pos]
if (pre[next] == pos):
pass
elif ((code[next] in (JF, JA)) and (target == self.get_target(next))):
if (code[pre[next]] == PJIF):
if ((code[next] == JF) or (target != rtarget) or (code[pre[pre[rtarget]]] not in (JA, RETURN_VALUE))):
self.fixed_jumps[pos] = pre[next]
return
elif ((code[next] == JA) and (code[target] in (JA, JF))):
next_target = self.get_target(next)
if (self.get_target(target) == next_target):
self.fixed_jumps[pos] = pre[next]
return
elif ((code[next_target] in (JA, JF)) and (self.get_target(next_target) == self.get_target(target))):
self.fixed_jumps[pos] = pre[next]
return
if (pos in self.ignore_if):
return
if ((code[pre[rtarget]] == JA) and (pre[rtarget] in self.stmts) and (pre[rtarget] != pos) and (pre[pre[rtarget]] != pos)):
if ((code[rtarget] == JA) and (code[(rtarget + 3)] == POP_BLOCK)):
if (code[pre[pre[rtarget]]] != JA):
pass
elif (self.get_target(pre[pre[rtarget]]) != target):
pass
else:
rtarget = pre[rtarget]
else:
rtarget = pre[rtarget]
if (code[pre[rtarget]] in (JA, JF)):
if_end = self.get_target(pre[rtarget])
if ((if_end < pre[rtarget]) and (code[pre[if_end]] == SETUP_LOOP)):
if (if_end > start):
return
end = self.restrict_to_parent(if_end, parent)
self.structs.append({'type': 'if-then', 'start': start, 'end': pre[rtarget]})
self.not_continue.add(pre[rtarget])
if (rtarget < end):
self.structs.append({'type': 'if-else', 'start': rtarget, 'end': end})
elif (code[pre[rtarget]] == RETURN_VALUE):
self.structs.append({'type': 'if-then', 'start': start, 'end': rtarget})
self.return_end_ifs.add(pre[rtarget])
elif (op in (JUMP_IF_FALSE_OR_POP, JUMP_IF_TRUE_OR_POP)):
target = self.get_target(pos, op)
self.fixed_jumps[pos] = self.restrict_to_parent(target, parent)
|
'Detect all offsets in a byte code which are jump targets.
Return the list of offsets.
This procedure is modelled after dis.findlables(), but here
for each target the number of jumps are counted.'
| def find_jump_targets(self, code):
| hasjrel = dis.hasjrel
hasjabs = dis.hasjabs
n = len(code)
self.structs = [{'type': 'root', 'start': 0, 'end': (n - 1)}]
self.loops = []
self.fixed_jumps = {}
self.ignore_if = set()
self.build_stmt_indices()
self.not_continue = set()
self.return_end_ifs = set()
targets = {}
for i in self.op_range(0, n):
op = code[i]
self.detect_structure(i, op)
if (op >= HAVE_ARGUMENT):
label = self.fixed_jumps.get(i)
oparg = (code[(i + 1)] + (code[(i + 2)] * 256))
if (label is None):
if ((op in hasjrel) and (op != FOR_ITER)):
label = ((i + 3) + oparg)
elif (op in hasjabs):
if (op in (JUMP_IF_FALSE_OR_POP, JUMP_IF_TRUE_OR_POP)):
if (oparg > i):
label = oparg
if ((label is not None) and (label != (-1))):
targets[label] = (targets.get(label, []) + [i])
elif ((op == END_FINALLY) and (i in self.fixed_jumps)):
label = self.fixed_jumps[i]
targets[label] = (targets.get(label, []) + [i])
return targets
|
'exec_stmt ::= expr exprlist DUP_TOP EXEC_STMT
exec_stmt ::= expr exprlist EXEC_STMT'
| def n_exec_stmt(self, node):
| self.write(self.indent, 'exec ')
self.preorder(node[0])
if (node[1][0] != NONE):
sep = ' in '
for subnode in node[1]:
self.write(sep)
sep = ', '
self.preorder(subnode)
self.print_()
self.prune()
|
'prettyprint a mapexpr
\'mapexpr\' is something like k = {\'a\': 1, \'b\': 42 }"'
| def n_mapexpr(self, node):
| p = self.prec
self.prec = 100
assert (node[(-1)] == 'kvlist')
node = node[(-1)]
self.indentMore(INDENT_PER_LEVEL)
line_seperator = (',\n' + self.indent)
sep = INDENT_PER_LEVEL[:(-1)]
self.write('{')
for kv in node:
assert (kv in ('kv', 'kv2', 'kv3'))
if (kv == 'kv'):
name = self.traverse(kv[(-2)], indent='')
value = self.traverse(kv[1], indent=(self.indent + ((len(name) + 2) * ' ')))
elif (kv == 'kv2'):
name = self.traverse(kv[1], indent='')
value = self.traverse(kv[(-3)], indent=(self.indent + ((len(name) + 2) * ' ')))
elif (kv == 'kv3'):
name = self.traverse(kv[(-2)], indent='')
value = self.traverse(kv[0], indent=(self.indent + ((len(name) + 2) * ' ')))
self.write(sep, name, ': ', value)
sep = line_seperator
self.write('}')
self.indentLess(INDENT_PER_LEVEL)
self.prec = p
self.prune()
|
'prettyprint a list or tuple'
| def n_build_list(self, node):
| p = self.prec
self.prec = 100
lastnode = node.pop()
lastnodetype = lastnode.type
if lastnodetype.startswith('BUILD_LIST'):
self.write('[')
endchar = ']'
elif lastnodetype.startswith('BUILD_TUPLE'):
self.write('(')
endchar = ')'
elif lastnodetype.startswith('BUILD_SET'):
self.write('{')
endchar = '}'
elif lastnodetype.startswith('ROT_TWO'):
self.write('(')
endchar = ')'
else:
raise 'Internal Error: n_build_list expects list or tuple'
flat_elems = []
for elem in node:
if (elem == 'expr1024'):
for subelem in elem:
for subsubelem in subelem:
flat_elems.append(subsubelem)
elif (elem == 'expr32'):
for subelem in elem:
flat_elems.append(subelem)
else:
flat_elems.append(elem)
self.indentMore(INDENT_PER_LEVEL)
if (lastnode.attr > 3):
line_separator = (',\n' + self.indent)
else:
line_separator = ', '
sep = INDENT_PER_LEVEL[:(-1)]
for elem in flat_elems:
if (elem == 'ROT_THREE'):
continue
assert (elem == 'expr')
value = self.traverse(elem)
self.write(sep, value)
sep = line_separator
if ((lastnode.attr == 1) and lastnodetype.startswith('BUILD_TUPLE')):
self.write(',')
self.write(endchar)
self.indentLess(INDENT_PER_LEVEL)
self.prec = p
self.prune()
|
'Special handling for opcodes that take a variable number
of arguments -- we add a new entry for each in TABLE_R.'
| def customize(self, customize):
| for (k, v) in customize.items():
if TABLE_R.has_key(k):
continue
op = k[:k.rfind('_')]
if (op == 'CALL_FUNCTION'):
TABLE_R[k] = ('%c(%P)', 0, (1, (-1), ', ', 100))
elif (op in ('CALL_FUNCTION_VAR', 'CALL_FUNCTION_VAR_KW', 'CALL_FUNCTION_KW')):
if (v == 0):
str = '%c(%C'
p2 = (0, 0, None)
else:
str = '%c(%C, '
p2 = (1, (-2), ', ')
if (op == 'CALL_FUNCTION_VAR'):
str += '*%c)'
entry = (str, 0, p2, (-2))
elif (op == 'CALL_FUNCTION_KW'):
str += '**%c)'
entry = (str, 0, p2, (-2))
else:
str += '*%c, **%c)'
if p2[2]:
p2 = (1, (-3), ', ')
entry = (str, 0, p2, (-3), (-2))
TABLE_R[k] = entry
|
'If the name of the formal parameter starts with dot,
it\'s a tuple parameter, like this:
# def MyFunc(xx, (a,b,c), yy):
# print a, b*2, c*42
In byte-code, the whole tuple is assigned to parameter \'.1\' and
then the tuple gets unpacked to \'a\', \'b\' and \'c\'.
Since identifiers starting with a dot are illegal in Python,
we can search for the byte-code equivalent to \'(a,b,c) = .1\''
| def get_tuple_parameter(self, ast, name):
| assert (ast == 'stmts')
for i in range(len(ast)):
assert (ast[i][0] == 'stmt')
node = ast[i][0][0]
if ((node == 'assign') and (node[0] == ASSIGN_TUPLE_PARAM(name))):
del ast[i]
assert (node[1] == 'designator')
return (('(' + self.traverse(node[1])) + ')')
raise ("Can't find tuple parameter" % name)
|
'Dump function defintion, doc string, and function body.'
| def make_function(self, node, isLambda, nested=1):
| def build_param(ast, name, default):
'build parameters:\n - handle defaults\n - handle format tuple parameters\n '
if name.startswith('.'):
name = self.get_tuple_parameter(ast, name)
if default:
if self.showast:
print '--', name
print default
print '--'
result = ('%s = %s' % (name, self.traverse(default, indent='')))
if (result[(-2):] == '= '):
result += 'None'
return result
else:
return name
defparams = node[:node[(-1)].attr]
code = node[(-2)].attr
assert (type(code) == CodeType)
code = Code(code, self.scanner, self.currentclass)
argc = code.co_argcount
paramnames = list(code.co_varnames[:argc])
paramnames.reverse()
defparams.reverse()
try:
ast = self.build_ast(code._tokens, code._customize, isLambda=isLambda, noneInNames=('None' in code.co_names))
except ParserError as p:
self.write(str(p))
self.ERROR = p
return
params = []
for (name, default) in map((lambda a, b: (a, b)), paramnames, defparams):
params.append(build_param(ast, name, default))
params.reverse()
if (4 & code.co_flags):
params.append(('*%s' % code.co_varnames[argc]))
argc += 1
if (8 & code.co_flags):
params.append(('**%s' % code.co_varnames[argc]))
argc += 1
indent = self.indent
if isLambda:
self.write('lambda ', ', '.join(params), ': ')
else:
self.print_('(', ', '.join(params), '):')
if ((len(code.co_consts) > 0) and (code.co_consts[0] != None)):
self.print_docstring(indent, code.co_consts[0])
code._tokens = None
assert (ast == 'stmts')
all_globals = find_all_globals(ast, set())
for g in ((all_globals & self.mod_globs) | find_globals(ast, set())):
self.print_(self.indent, 'global ', g)
self.mod_globs -= all_globals
rn = (('None' in code.co_names) and (not find_none(ast)))
self.gen_source(ast, code._customize, isLambda=isLambda, returnNone=rn)
code._tokens = None
code._customize = None
|
'Dump class definition, doc string and class body.'
| def build_class(self, code):
| assert (type(code) == CodeType)
code = Code(code, self.scanner, self.currentclass)
indent = self.indent
ast = self.build_ast(code._tokens, code._customize)
code._tokens = None
assert (ast == 'stmts')
if (ast[0][0] == NAME_MODULE):
del ast[0]
if (code.co_consts and (code.co_consts[0] != None) and (ast[0][0] == ASSIGN_DOC_STRING(code.co_consts[0]))):
self.print_docstring(indent, code.co_consts[0])
self.print_()
del ast[0]
if (ast[(-1)][0] == RETURN_LOCALS):
del ast[(-1)]
for g in find_globals(ast, set()):
self.print_(indent, 'global ', g)
self.gen_source(ast, code._customize)
code._tokens = None
code._customize = None
|
'convert AST to source code'
| def gen_source(self, ast, customize, isLambda=0, returnNone=False):
| rn = self.return_none
self.return_none = returnNone
if (len(ast) == 0):
self.print_(self.indent, 'pass')
else:
self.customize(customize)
if isLambda:
self.write(self.traverse(ast, isLambda=isLambda))
else:
self.print_(self.traverse(ast, isLambda=isLambda))
self.return_none = rn
|
'Places new Gmail notifications in the Notifier\'s queue.'
| def handleEmailNotifications(self, lastDate):
| emails = Gmail.fetchUnreadEmails(self.profile, since=lastDate)
if emails:
lastDate = Gmail.getMostRecentDate(emails)
def styleEmail(e):
return ('New email from %s.' % Gmail.getSender(e))
for e in emails:
self.q.put(styleEmail(e))
return lastDate
|
'Returns a notification. Note that this function is consuming.'
| def getNotification(self):
| try:
notif = self.q.get(block=False)
return notif
except Queue.Empty:
return None
|
'Return a list of notifications in chronological order.
Note that this function is consuming, so consecutive calls
will yield different results.'
| def getAllNotifications(self):
| notifs = []
notif = self.getNotification()
while notif:
notifs.append(notif)
notif = self.getNotification()
return notifs
|
'Calculates a revision from phrases by using the SHA1 hash function.
Arguments:
phrases -- a list of phrases
Returns:
A revision string for given phrases.'
| @classmethod
def phrases_to_revision(cls, phrases):
| sorted_phrases = sorted(phrases)
joined_phrases = '\n'.join(sorted_phrases)
sha1 = hashlib.sha1()
sha1.update(joined_phrases)
return sha1.hexdigest()
|
'Initializes a new Vocabulary instance.
Optional Arguments:
name -- (optional) the name of the vocabulary (Default: \'default\')
path -- (optional) the path in which the vocabulary exists or will
be created (Default: \'.\')'
| def __init__(self, name='default', path='.'):
| self.name = name
self.path = os.path.abspath(os.path.join(path, self.PATH_PREFIX, name))
self._logger = logging.getLogger(__name__)
|
'Returns:
The path of the the revision file as string'
| @property
def revision_file(self):
| return os.path.join(self.path, 'revision')
|
'Checks if the vocabulary is compiled by checking if the revision file
is readable. This method should be overridden by subclasses to check
for class-specific additional files, too.
Returns:
True if the dictionary is compiled, else False'
| @abstractproperty
def is_compiled(self):
| return os.access(self.revision_file, os.R_OK)
|
'Reads the compiled revision from the revision file.
Returns:
the revision of this vocabulary (i.e. the string
inside the revision file), or None if is_compiled
if False'
| @property
def compiled_revision(self):
| if (not self.is_compiled):
return None
with open(self.revision_file, 'r') as f:
revision = f.read().strip()
self._logger.debug("compiled_revision is '%s'", revision)
return revision
|
'Convenience method to check if this vocabulary exactly contains the
phrases passed to this method.
Arguments:
phrases -- a list of phrases
Returns:
True if phrases exactly matches the phrases inside this
vocabulary.'
| def matches_phrases(self, phrases):
| return (self.compiled_revision == self.phrases_to_revision(phrases))
|
'Compiles this vocabulary. If the force argument is True, compilation
will be forced regardless of necessity (which means that the
preliminary check if the current revision already equals the
revision after compilation will be skipped).
This method is not meant to be overridden by subclasses - use the
_compile_vocabulary()-method instead.
Arguments:
phrases -- a list of phrases that this vocabulary will contain
force -- (optional) forces compilation (Default: False)
Returns:
The revision of the compiled vocabulary'
| def compile(self, phrases, force=False):
| revision = self.phrases_to_revision(phrases)
if ((not force) and (self.compiled_revision == revision)):
self._logger.debug(('Compilation not neccessary, compiled ' + 'version matches phrases.'))
return revision
if (not os.path.exists(self.path)):
self._logger.debug(("Vocabulary dir '%s' does not exist, " + 'creating...'), self.path)
try:
os.makedirs(self.path)
except OSError:
self._logger.error("Couldn't create vocabulary dir '%s'", self.path, exc_info=True)
raise
try:
with open(self.revision_file, 'w') as f:
f.write(revision)
except (OSError, IOError):
self._logger.error("Couldn't write revision file in '%s'", self.revision_file, exc_info=True)
raise
else:
self._logger.info('Starting compilation...')
try:
self._compile_vocabulary(phrases)
except Exception as e:
self._logger.error(('Fatal compilation Error occured, ' + 'cleaning up...'), exc_info=True)
try:
os.remove(self.revision_file)
except OSError:
pass
raise e
else:
self._logger.info('Compilation done.')
return revision
|
'Checks if the vocabulary is compiled by checking if the revision
file is readable.
Returns:
True if this vocabulary has been compiled, else False'
| @property
def is_compiled(self):
| return super(self.__class__, self).is_compiled
|
'Does nothing (because this is a dummy class for testing purposes).'
| def _compile_vocabulary(self, phrases):
| pass
|
'Returns:
The path of the the pocketsphinx languagemodel file as string'
| @property
def languagemodel_file(self):
| return os.path.join(self.path, 'languagemodel')
|
'Returns:
The path of the pocketsphinx dictionary file as string'
| @property
def dictionary_file(self):
| return os.path.join(self.path, 'dictionary')
|
'Checks if the vocabulary is compiled by checking if the revision,
languagemodel and dictionary files are readable.
Returns:
True if this vocabulary has been compiled, else False'
| @property
def is_compiled(self):
| return (super(self.__class__, self).is_compiled and os.access(self.languagemodel_file, os.R_OK) and os.access(self.dictionary_file, os.R_OK))
|
'Convenience property to use this Vocabulary with the __init__() method
of the pocketsphinx.Decoder class.
Returns:
A dict containing kwargs for the pocketsphinx.Decoder.__init__()
method.
Example:
decoder = pocketsphinx.Decoder(**vocab_instance.decoder_kwargs,
hmm=\'/path/to/hmm\')'
| @property
def decoder_kwargs(self):
| return {'lm': self.languagemodel_file, 'dict': self.dictionary_file}
|
'Compiles the vocabulary to the Pocketsphinx format by creating a
languagemodel and a dictionary.
Arguments:
phrases -- a list of phrases that this vocabulary will contain'
| def _compile_vocabulary(self, phrases):
| text = ' '.join([('<s> %s </s>' % phrase) for phrase in phrases])
self._logger.debug('Compiling languagemodel...')
vocabulary = self._compile_languagemodel(text, self.languagemodel_file)
self._logger.debug('Starting dictionary...')
self._compile_dictionary(vocabulary, self.dictionary_file)
|
'Compiles the languagemodel from a text.
Arguments:
text -- the text the languagemodel will be generated from
output_file -- the path of the file this languagemodel will
be written to
Returns:
A list of all unique words this vocabulary contains.'
| def _compile_languagemodel(self, text, output_file):
| with tempfile.NamedTemporaryFile(suffix='.vocab', delete=False) as f:
vocab_file = f.name
self._logger.debug("Creating vocab file: '%s'", vocab_file)
cmuclmtk.text2vocab(text, vocab_file)
self._logger.debug("Creating languagemodel file: '%s'", output_file)
cmuclmtk.text2lm(text, output_file, vocab_file=vocab_file)
self._logger.debug(('Getting words from vocab file and removing it ' + 'afterwards...'))
words = []
with open(vocab_file, 'r') as f:
for line in f:
line = line.strip()
if ((not line.startswith('#')) and (line not in ('<s>', '</s>'))):
words.append(line)
os.remove(vocab_file)
return words
|
'Compiles the dictionary from a list of words.
Arguments:
words -- a list of all unique words this vocabulary contains
output_file -- the path of the file this dictionary will
be written to'
| def _compile_dictionary(self, words, output_file):
| self._logger.debug('Getting phonemes for %d words...', len(words))
g2pconverter = PhonetisaurusG2P(**PhonetisaurusG2P.get_config())
phonemes = g2pconverter.translate(words)
self._logger.debug("Creating dict file: '%s'", output_file)
with open(output_file, 'w') as f:
for (word, pronounciations) in phonemes.items():
for (i, pronounciation) in enumerate(pronounciations, start=1):
if (i == 1):
line = ('%s DCTB %s\n' % (word, pronounciation))
else:
line = ('%s(%d) DCTB %s\n' % (word, i, pronounciation))
f.write(line)
|
'Returns:
The path of the the julius dfa file as string'
| @property
def dfa_file(self):
| return os.path.join(self.path, 'dfa')
|
'Returns:
The path of the the julius dict file as string'
| @property
def dict_file(self):
| return os.path.join(self.path, 'dict')
|
'Prepare the client and music variables'
| def __init__(self, server='localhost', port=6600):
| self.server = server
self.port = port
self.client = mpd.MPDClient()
self.client.timeout = None
self.client.idletimeout = None
self.client.connect(self.server, self.port)
self.playlists = [x['playlist'] for x in self.client.listplaylists()]
self.client.clear()
for playlist in self.playlists:
self.client.load(playlist)
self.songs = []
self.song_titles = []
self.song_artists = []
soup = self.client.playlist()
for i in range(0, (len(soup) / 10)):
index = (i * 10)
id = soup[index].strip()
title = soup[(index + 3)].strip().upper()
artist = soup[(index + 2)].strip().upper()
album = soup[(index + 4)].strip().upper()
self.songs.append(Song(id, title, artist, album))
self.song_titles.append(title)
self.song_artists.append(artist)
|
'Plays the current song or accepts a song to play.
Arguments:
songs -- a list of song objects
playlist_name -- user-defined, something like "Love Song Playlist"'
| @reconnect
def play(self, songs=False, playlist_name=False):
| if songs:
self.client.clear()
for song in songs:
try:
self.client.add(song.id)
except:
pass
if playlist_name:
self.client.clear()
self.client.load(playlist_name)
self.client.play()
|
'Returns the list of unique words that comprise song and artist titles'
| def get_soup(self):
| soup = []
for song in self.songs:
song_words = song.title.split(' ')
artist_words = song.artist.split(' ')
soup.extend(song_words)
soup.extend(artist_words)
title_trans = ''.join(((chr(c) if (chr(c).isupper() or chr(c).islower()) else '_') for c in range(256)))
soup = [x.decode('utf-8').encode('ascii', 'ignore').upper().translate(title_trans).replace('_', '') for x in soup]
soup = [x for x in soup if (x != '')]
return list(set(soup))
|
'Returns the list of unique words that comprise playlist names'
| def get_soup_playlist(self):
| soup = []
for name in self.playlists:
soup.extend(name.split(' '))
title_trans = ''.join(((chr(c) if (chr(c).isupper() or chr(c).islower()) else '_') for c in range(256)))
soup = [x.decode('utf-8').encode('ascii', 'ignore').upper().translate(title_trans).replace('_', '') for x in soup]
soup = [x for x in soup if (x != '')]
return list(set(soup))
|
'Returns the list of PHRASES that comprise song and artist titles'
| def get_soup_separated(self):
| title_soup = [song.title for song in self.songs]
artist_soup = [song.artist for song in self.songs]
soup = list(set((title_soup + artist_soup)))
title_trans = ''.join(((chr(c) if (chr(c).isupper() or chr(c).islower()) else '_') for c in range(256)))
soup = [x.decode('utf-8').encode('ascii', 'ignore').upper().translate(title_trans).replace('_', ' ') for x in soup]
soup = [re.sub(' +', ' ', x) for x in soup if (x != '')]
return soup
|
'Returns songs matching a query best as possible on either artist
field, etc'
| def fuzzy_songs(self, query):
| query = query.upper()
matched_song_titles = difflib.get_close_matches(query, self.song_titles)
matched_song_artists = difflib.get_close_matches(query, self.song_artists)
strict_priority_title = [x for x in matched_song_titles if (x == query)]
strict_priority_artists = [x for x in matched_song_artists if (x == query)]
if strict_priority_title:
matched_song_titles = strict_priority_title
if strict_priority_artists:
matched_song_artists = strict_priority_artists
matched_songs_bytitle = [song for song in self.songs if (song.title in matched_song_titles)]
matched_songs_byartist = [song for song in self.songs if (song.artist in matched_song_artists)]
matches = list(set((matched_songs_bytitle + matched_songs_byartist)))
return matches
|
'returns playlist names that match query best as possible'
| def fuzzy_playlists(self, query):
| query = query.upper()
lookup = {n.upper(): n for n in self.playlists}
results = [lookup[r] for r in difflib.get_close_matches(query, lookup)]
return results
|
'Initiates the pocketsphinx instance.
Arguments:
speaker -- handles platform-independent audio output
passive_stt_engine -- performs STT while Jasper is in passive listen
mode
acive_stt_engine -- performs STT while Jasper is in active listen mode'
| def __init__(self, speaker, passive_stt_engine, active_stt_engine):
| self._logger = logging.getLogger(__name__)
self.speaker = speaker
self.passive_stt_engine = passive_stt_engine
self.active_stt_engine = active_stt_engine
self._logger.info((('Initializing PyAudio. ALSA/Jack error messages ' + 'that pop up during this process are normal and ') + 'can usually be safely ignored.'))
self._audio = pyaudio.PyAudio()
self._logger.info('Initialization of PyAudio completed.')
|
'Listens for PERSONA in everyday sound. Times out after LISTEN_TIME, so
needs to be restarted.'
| def passiveListen(self, PERSONA):
| THRESHOLD_MULTIPLIER = 1.8
RATE = 16000
CHUNK = 1024
THRESHOLD_TIME = 1
LISTEN_TIME = 10
stream = self._audio.open(format=pyaudio.paInt16, channels=1, rate=RATE, input=True, frames_per_buffer=CHUNK)
frames = []
lastN = [i for i in range(30)]
for i in range(0, ((RATE / CHUNK) * THRESHOLD_TIME)):
data = stream.read(CHUNK)
frames.append(data)
lastN.pop(0)
lastN.append(self.getScore(data))
average = (sum(lastN) / len(lastN))
THRESHOLD = (average * THRESHOLD_MULTIPLIER)
frames = []
didDetect = False
for i in range(0, ((RATE / CHUNK) * LISTEN_TIME)):
data = stream.read(CHUNK)
frames.append(data)
score = self.getScore(data)
if (score > THRESHOLD):
didDetect = True
break
if (not didDetect):
print 'No disturbance detected'
stream.stop_stream()
stream.close()
return (None, None)
frames = frames[(-20):]
DELAY_MULTIPLIER = 1
for i in range(0, ((RATE / CHUNK) * DELAY_MULTIPLIER)):
data = stream.read(CHUNK)
frames.append(data)
stream.stop_stream()
stream.close()
with tempfile.NamedTemporaryFile(mode='w+b') as f:
wav_fp = wave.open(f, 'wb')
wav_fp.setnchannels(1)
wav_fp.setsampwidth(pyaudio.get_sample_size(pyaudio.paInt16))
wav_fp.setframerate(RATE)
wav_fp.writeframes(''.join(frames))
wav_fp.close()
f.seek(0)
transcribed = self.passive_stt_engine.transcribe(f)
if any(((PERSONA in phrase) for phrase in transcribed)):
return (THRESHOLD, PERSONA)
return (False, transcribed)
|
'Records until a second of silence or times out after 12 seconds
Returns the first matching string or None'
| def activeListen(self, THRESHOLD=None, LISTEN=True, MUSIC=False):
| options = self.activeListenToAllOptions(THRESHOLD, LISTEN, MUSIC)
if options:
return options[0]
|
'Records until a second of silence or times out after 12 seconds
Returns a list of the matching options or None'
| def activeListenToAllOptions(self, THRESHOLD=None, LISTEN=True, MUSIC=False):
| RATE = 16000
CHUNK = 1024
LISTEN_TIME = 12
if (THRESHOLD is None):
THRESHOLD = self.fetchThreshold()
self.speaker.play(jasperpath.data('audio', 'beep_hi.wav'))
stream = self._audio.open(format=pyaudio.paInt16, channels=1, rate=RATE, input=True, frames_per_buffer=CHUNK)
frames = []
lastN = [(THRESHOLD * 1.2) for i in range(30)]
for i in range(0, ((RATE / CHUNK) * LISTEN_TIME)):
data = stream.read(CHUNK)
frames.append(data)
score = self.getScore(data)
lastN.pop(0)
lastN.append(score)
average = (sum(lastN) / float(len(lastN)))
if (average < (THRESHOLD * 0.8)):
break
self.speaker.play(jasperpath.data('audio', 'beep_lo.wav'))
stream.stop_stream()
stream.close()
with tempfile.SpooledTemporaryFile(mode='w+b') as f:
wav_fp = wave.open(f, 'wb')
wav_fp.setnchannels(1)
wav_fp.setsampwidth(pyaudio.get_sample_size(pyaudio.paInt16))
wav_fp.setframerate(RATE)
wav_fp.writeframes(''.join(frames))
wav_fp.close()
f.seek(0)
return self.active_stt_engine.transcribe(f)
|
'Delegates user input to the handling function when activated.'
| def handleForever(self):
| self._logger.info("Starting to handle conversation with keyword '%s'.", self.persona)
while True:
notifications = self.notifier.getAllNotifications()
for notif in notifications:
self._logger.info("Received notification: '%s'", str(notif))
self._logger.debug("Started listening for keyword '%s'", self.persona)
(threshold, transcribed) = self.mic.passiveListen(self.persona)
self._logger.debug("Stopped listening for keyword '%s'", self.persona)
if ((not transcribed) or (not threshold)):
self._logger.info('Nothing has been said or transcribed.')
continue
self._logger.info("Keyword '%s' has been said!", self.persona)
self._logger.debug('Started to listen actively with threshold: %r', threshold)
input = self.mic.activeListenToAllOptions(threshold)
self._logger.debug('Stopped to listen actively with threshold: %r', threshold)
if input:
self.brain.query(input)
else:
self.mic.say('Pardon?')
|
'Instantiates a new Brain object, which cross-references user
input with a list of modules. Note that the order of brain.modules
matters, as the Brain will cease execution on the first module
that accepts a given input.
Arguments:
mic -- used to interact with the user (for both input and output)
profile -- contains information related to the user (e.g., phone
number)'
| def __init__(self, mic, profile):
| self.mic = mic
self.profile = profile
self.modules = self.get_modules()
self._logger = logging.getLogger(__name__)
|
'Dynamically loads all the modules in the modules folder and sorts
them by the PRIORITY key. If no PRIORITY is defined for a given
module, a priority of 0 is assumed.'
| @classmethod
def get_modules(cls):
| logger = logging.getLogger(__name__)
locations = [jasperpath.PLUGIN_PATH]
logger.debug('Looking for modules in: %s', ', '.join([("'%s'" % location) for location in locations]))
modules = []
for (finder, name, ispkg) in pkgutil.walk_packages(locations):
try:
loader = finder.find_module(name)
mod = loader.load_module(name)
except:
logger.warning("Skipped module '%s' due to an error.", name, exc_info=True)
else:
if hasattr(mod, 'WORDS'):
logger.debug("Found module '%s' with words: %r", name, mod.WORDS)
modules.append(mod)
else:
logger.warning(("Skipped module '%s' because it misses " + 'the WORDS constant.'), name)
modules.sort(key=(lambda mod: (mod.PRIORITY if hasattr(mod, 'PRIORITY') else 0)), reverse=True)
return modules
|
'Passes user input to the appropriate module, testing it against
each candidate module\'s isValid function.
Arguments:
text -- user input, typically speech, to be parsed by a module'
| def query(self, texts):
| for module in self.modules:
for text in texts:
if module.isValid(text):
self._logger.debug(("'%s' is a valid phrase for module " + "'%s'"), text, module.__name__)
try:
module.handle(text, self.mic, self.profile)
except Exception:
self._logger.error('Failed to execute module', exc_info=True)
self.mic.say(("I'm sorry. I had some trouble with " + 'that operation. Please try again later.'))
else:
self._logger.debug(("Handling of phrase '%s' by " + "module '%s' completed"), text, module.__name__)
finally:
return
self._logger.debug(('No module was able to handle any of these ' + 'phrases: %r'), texts)
|
'Initiates the pocketsphinx instance.
Arguments:
vocabulary -- a PocketsphinxVocabulary instance
hmm_dir -- the path of the Hidden Markov Model (HMM)'
| def __init__(self, vocabulary, hmm_dir=('/usr/local/share/' + 'pocketsphinx/model/hmm/en_US/hub4wsj_sc_8k')):
| self._logger = logging.getLogger(__name__)
try:
import pocketsphinx as ps
except:
import pocketsphinx as ps
with tempfile.NamedTemporaryFile(prefix='psdecoder_', suffix='.log', delete=False) as f:
self._logfile = f.name
self._logger.debug(('Initializing PocketSphinx Decoder with hmm_dir ' + "'%s'"), hmm_dir)
if (not os.path.exists(hmm_dir)):
msg = (("hmm_dir '%s' does not exist! Please make sure that you " + 'have set the correct hmm_dir in your profile.') % hmm_dir)
self._logger.error(msg)
raise RuntimeError(msg)
missing_hmm_files = []
for fname in ('mdef', 'feat.params', 'means', 'noisedict', 'transition_matrices', 'variances'):
if (not os.path.exists(os.path.join(hmm_dir, fname))):
missing_hmm_files.append(fname)
mixweights = os.path.exists(os.path.join(hmm_dir, 'mixture_weights'))
sendump = os.path.exists(os.path.join(hmm_dir, 'sendump'))
if ((not mixweights) and (not sendump)):
missing_hmm_files.append('mixture_weights or sendump')
if missing_hmm_files:
self._logger.warning((("hmm_dir '%s' is missing files: %s. Please " + 'make sure that you have set the correct ') + 'hmm_dir in your profile.'), hmm_dir, ', '.join(missing_hmm_files))
self._decoder = ps.Decoder(hmm=hmm_dir, logfn=self._logfile, **vocabulary.decoder_kwargs)
|
'Performs STT, transcribing an audio file and returning the result.
Arguments:
fp -- a file object containing audio data'
| def transcribe(self, fp):
| fp.seek(44)
data = fp.read()
self._decoder.start_utt()
self._decoder.process_raw(data, False, True)
self._decoder.end_utt()
result = self._decoder.get_hyp()
with open(self._logfile, 'r+') as f:
for line in f:
self._logger.debug(line.strip())
f.truncate()
transcribed = [result[0]]
self._logger.info('Transcribed: %r', transcribed)
return transcribed
|
'Arguments:
api_key - the public api key which allows access to Google APIs'
| def __init__(self, api_key=None, language='en-us'):
| self._logger = logging.getLogger(__name__)
self._request_url = None
self._language = None
self._api_key = None
self._http = requests.Session()
self.language = language
self.api_key = api_key
|
'Performs STT via the Google Speech API, transcribing an audio file and
returning an English string.
Arguments:
audio_file_path -- the path to the .wav file to be transcribed'
| def transcribe(self, fp):
| if (not self.api_key):
self._logger.critical(('API key missing, transcription request ' + 'aborted.'))
return []
elif (not self.language):
self._logger.critical(('Language info missing, transcription ' + 'request aborted.'))
return []
wav = wave.open(fp, 'rb')
frame_rate = wav.getframerate()
wav.close()
data = fp.read()
headers = {'content-type': ('audio/l16; rate=%s' % frame_rate)}
r = self._http.post(self.request_url, data=data, headers=headers)
try:
r.raise_for_status()
except requests.exceptions.HTTPError:
self._logger.critical('Request failed with http status %d', r.status_code)
if (r.status_code == requests.codes['forbidden']):
self._logger.warning(('Status 403 is probably caused by an ' + 'invalid Google API key.'))
return []
r.encoding = 'utf-8'
try:
response = json.loads(list(r.text.strip().split('\n', 1))[(-1)])
if (len(response['result']) == 0):
raise ValueError('Nothing has been transcribed.')
results = [alt['transcript'] for alt in response['result'][0]['alternative']]
except ValueError as e:
self._logger.warning('Empty response: %s', e.args[0])
results = []
except (KeyError, IndexError):
self._logger.warning('Cannot parse response.', exc_info=True)
results = []
else:
results = tuple((result.upper() for result in results))
self._logger.info('Transcribed: %r', results)
return results
|
'Does Jasper recognize his name (i.e., passive listen)?'
| def testTranscribeJasper(self):
| with open(self.jasper_clip, mode='rb') as f:
transcription = self.passive_stt_engine.transcribe(f)
self.assertIn('JASPER', transcription)
|
'Does Jasper recognize \'time\' (i.e., active listen)?'
| def testTranscribe(self):
| with open(self.time_clip, mode='rb') as f:
transcription = self.active_stt_engine.transcribe(f)
self.assertIn('TIME', transcription)
|
'Does Brain correctly log errors when raised by modules?'
| def testLog(self):
| my_brain = TestBrain._emptyBrain()
unclear = my_brain.modules[(-1)]
with mock.patch.object(unclear, 'handle') as mocked_handle:
with mock.patch.object(my_brain._logger, 'error') as mocked_log:
mocked_handle.side_effect = KeyError('foo')
my_brain.query('zzz gibberish zzz')
self.assertTrue(mocked_log.called)
|
'Does Brain sort modules by priority?'
| def testSortByPriority(self):
| my_brain = TestBrain._emptyBrain()
priorities = filter((lambda m: hasattr(m, 'PRIORITY')), my_brain.modules)
target = sorted(priorities, key=(lambda m: m.PRIORITY), reverse=True)
self.assertEqual(target, priorities)
|
'Does Brain correctly send query to higher-priority module?'
| def testPriority(self):
| my_brain = TestBrain._emptyBrain()
hn_module = 'HN'
hn = filter((lambda m: (m.__name__ == hn_module)), my_brain.modules)[0]
with mock.patch.object(hn, 'handle') as mocked_handle:
my_brain.query(['hacker news'])
self.assertTrue(mocked_handle.called)
|
'Generic method for spoofing conversation.
Arguments:
query -- The initial input to the server.
inputs -- Additional input, if conversation is extended.
Returns:
The server\'s responses, in a list.'
| def runConversation(self, query, inputs, module):
| self.assertTrue(module.isValid(query))
mic = test_mic.Mic(inputs)
module.handle(query, mic, self.profile)
return mic.outputs
|
'The lines below is a spider contract. For more info see:
http://doc.scrapy.org/en/latest/topics/contracts.html
@url http://www.dmoz.org/Computers/Programming/Languages/Python/Resources/
@scrapes name'
| def parse(self, response):
| sites = response.css('#site-list-content > div.site-item > div.title-and-desc')
items = []
for site in sites:
item = Website()
item['name'] = site.css('a > div.site-title::text').extract_first().strip()
item['url'] = site.xpath('a/@href').extract_first().strip()
item['description'] = site.css('div.site-descr::text').extract_first().strip()
items.append(item)
return items
|
'Since the original function always creates the directory, to resolve
the issue a new function had to be created. It\'s a simple copy and
was reduced for this case.
More info at:
https://github.com/scrapy/scrapy/pull/2005'
| def _copytree(self, src, dst):
| ignore = IGNORE
names = os.listdir(src)
ignored_names = ignore(src, names)
if (not os.path.exists(dst)):
os.makedirs(dst)
for name in names:
if (name in ignored_names):
continue
srcname = os.path.join(src, name)
dstname = os.path.join(dst, name)
if os.path.isdir(srcname):
self._copytree(srcname, dstname)
else:
copy2(srcname, dstname)
copystat(src, dst)
|
'Command syntax (preferably one-line). Do not include command name.'
| def syntax(self):
| return ''
|
'A short description of the command'
| def short_desc(self):
| return ''
|
'A long description of the command. Return short description when not
available. It cannot contain newlines, since contents will be formatted
by optparser which removes newlines and wraps text.'
| def long_desc(self):
| return self.short_desc()
|
'An extensive help for the command. It will be shown when using the
"help" command. It can contain newlines, since not post-formatting will
be applied to its contents.'
| def help(self):
| return self.long_desc()
|
'Populate option parse with options available for this command'
| def add_options(self, parser):
| group = OptionGroup(parser, 'Global Options')
group.add_option('--logfile', metavar='FILE', help='log file. if omitted stderr will be used')
group.add_option('-L', '--loglevel', metavar='LEVEL', default=None, help=('log level (default: %s)' % self.settings['LOG_LEVEL']))
group.add_option('--nolog', action='store_true', help='disable logging completely')
group.add_option('--profile', metavar='FILE', default=None, help='write python cProfile stats to FILE')
group.add_option('--pidfile', metavar='FILE', help='write process ID to FILE')
group.add_option('-s', '--set', action='append', default=[], metavar='NAME=VALUE', help='set/override setting (may be repeated)')
group.add_option('--pdb', action='store_true', help='enable pdb on failure')
parser.add_option_group(group)
|
'Entry point for running commands'
| def run(self, args, opts):
| raise NotImplementedError
|
'Generate the spider module, based on the given template'
| def _genspider(self, module, name, domain, template_name, template_file):
| tvars = {'project_name': self.settings.get('BOT_NAME'), 'ProjectName': string_camelcase(self.settings.get('BOT_NAME')), 'module': module, 'name': name, 'domain': domain, 'classname': ('%sSpider' % ''.join((s.capitalize() for s in module.split('_'))))}
if self.settings.get('NEWSPIDER_MODULE'):
spiders_module = import_module(self.settings['NEWSPIDER_MODULE'])
spiders_dir = abspath(dirname(spiders_module.__file__))
else:
spiders_module = None
spiders_dir = '.'
spider_file = ('%s.py' % join(spiders_dir, module))
shutil.copyfile(template_file, spider_file)
render_templatefile(spider_file, **tvars)
print(('Created spider %r using template %r ' % (name, template_name)), end=('' if spiders_module else '\n'))
if spiders_module:
print(('in module:\n %s.%s' % (spiders_module.__name__, module)))
|
'You can use this function to update the Scrapy objects that will be
available in the shell'
| def update_vars(self, vars):
| pass
|
'Start the execution engine'
| @defer.inlineCallbacks
def start(self):
| assert (not self.running), 'Engine already running'
self.start_time = time()
(yield self.signals.send_catch_log_deferred(signal=signals.engine_started))
self.running = True
self._closewait = defer.Deferred()
(yield self._closewait)
|
'Stop the execution engine gracefully'
| def stop(self):
| assert self.running, 'Engine not running'
self.running = False
dfd = self._close_all_spiders()
return dfd.addBoth((lambda _: self._finish_stopping_engine()))
|
'Close the execution engine gracefully.
If it has already been started, stop it. In all cases, close all spiders
and the downloader.'
| def close(self):
| if self.running:
return self.stop()
elif self.open_spiders:
return self._close_all_spiders()
else:
return defer.succeed(self.downloader.close())
|
'Pause the execution engine'
| def pause(self):
| self.paused = True
|
'Resume the execution engine'
| def unpause(self):
| self.paused = False
|
'Does the engine have capacity to handle more spiders'
| def has_capacity(self):
| return (not bool(self.slot))
|
'Called when a spider gets idle. This function is called when there
are no remaining pages to download or schedule. It can be called
multiple times. If some extension raises a DontCloseSpider exception
(in the spider_idle signal handler) the spider is not closed until the
next loop and this function is guaranteed to be called (at least) once
again for this spider.'
| def _spider_idle(self, spider):
| res = self.signals.send_catch_log(signal=signals.spider_idle, spider=spider, dont_log=DontCloseSpider)
if any(((isinstance(x, Failure) and isinstance(x.value, DontCloseSpider)) for (_, x) in res)):
return
if self.spider_is_idle(spider):
self.close_spider(spider, reason='finished')
|
'Close (cancel) spider and clear all its outstanding requests'
| def close_spider(self, spider, reason='cancelled'):
| slot = self.slot
if slot.closing:
return slot.closing
logger.info('Closing spider (%(reason)s)', {'reason': reason}, extra={'spider': spider})
dfd = slot.close()
def log_failure(msg):
def errback(failure):
logger.error(msg, exc_info=failure_to_exc_info(failure), extra={'spider': spider})
return errback
dfd.addBoth((lambda _: self.downloader.close()))
dfd.addErrback(log_failure('Downloader close failure'))
dfd.addBoth((lambda _: self.scraper.close_spider(spider)))
dfd.addErrback(log_failure('Scraper close failure'))
dfd.addBoth((lambda _: slot.scheduler.close(reason)))
dfd.addErrback(log_failure('Scheduler close failure'))
dfd.addBoth((lambda _: self.signals.send_catch_log_deferred(signal=signals.spider_closed, spider=spider, reason=reason)))
dfd.addErrback(log_failure('Error while sending spider_close signal'))
dfd.addBoth((lambda _: self.crawler.stats.close_spider(spider, reason=reason)))
dfd.addErrback(log_failure('Stats close failure'))
dfd.addBoth((lambda _: logger.info('Spider closed (%(reason)s)', {'reason': reason}, extra={'spider': spider})))
dfd.addBoth((lambda _: setattr(self, 'slot', None)))
dfd.addErrback(log_failure('Error while unassigning slot'))
dfd.addBoth((lambda _: setattr(self, 'spider', None)))
dfd.addErrback(log_failure('Error while unassigning spider'))
dfd.addBoth((lambda _: self._spider_closed_callback(spider)))
return dfd
|
'Open the given spider for scraping and allocate resources for it'
| @defer.inlineCallbacks
def open_spider(self, spider):
| self.slot = Slot()
(yield self.itemproc.open_spider(spider))
|
'Close a spider being scraped and release its resources'
| def close_spider(self, spider):
| slot = self.slot
slot.closing = defer.Deferred()
slot.closing.addCallback(self.itemproc.close_spider)
self._check_if_closing(spider, slot)
return slot.closing
|
'Return True if there isn\'t any more spiders to process'
| def is_idle(self):
| return (not self.slot)
|
'Handle the downloaded response or failure through the spider
callback/errback'
| def _scrape(self, response, request, spider):
| assert isinstance(response, (Response, Failure))
dfd = self._scrape2(response, request, spider)
dfd.addErrback(self.handle_spider_error, request, response, spider)
dfd.addCallback(self.handle_spider_output, request, response, spider)
return dfd
|
'Handle the different cases of request\'s result been a Response or a
Failure'
| def _scrape2(self, request_result, request, spider):
| if (not isinstance(request_result, Failure)):
return self.spidermw.scrape_response(self.call_spider, request_result, request, spider)
else:
dfd = self.call_spider(request_result, request, spider)
return dfd.addErrback(self._log_download_errors, request_result, request, spider)
|
'Process each Request/Item (given in the output parameter) returned
from the given spider'
| def _process_spidermw_output(self, output, request, response, spider):
| if isinstance(output, Request):
self.crawler.engine.crawl(request=output, spider=spider)
elif isinstance(output, (BaseItem, dict)):
self.slot.itemproc_size += 1
dfd = self.itemproc.process_item(output, spider)
dfd.addBoth(self._itemproc_finished, output, response, spider)
return dfd
elif (output is None):
pass
else:
typename = type(output).__name__
logger.error('Spider must return Request, BaseItem, dict or None, got %(typename)r in %(request)s', {'request': request, 'typename': typename}, extra={'spider': spider})
|
'Log and silence errors that come from the engine (typically download
errors that got propagated thru here)'
| def _log_download_errors(self, spider_failure, download_failure, request, spider):
| if (isinstance(download_failure, Failure) and (not download_failure.check(IgnoreRequest))):
if download_failure.frames:
logger.error('Error downloading %(request)s', {'request': request}, exc_info=failure_to_exc_info(download_failure), extra={'spider': spider})
else:
errmsg = download_failure.getErrorMessage()
if errmsg:
logger.error('Error downloading %(request)s: %(errmsg)s', {'request': request, 'errmsg': errmsg}, extra={'spider': spider})
if (spider_failure is not download_failure):
return spider_failure
|
'ItemProcessor finished for the given ``item`` and returned ``output``'
| def _itemproc_finished(self, output, item, response, spider):
| self.slot.itemproc_size -= 1
if isinstance(output, Failure):
ex = output.value
if isinstance(ex, DropItem):
logkws = self.logformatter.dropped(item, ex, response, spider)
logger.log(extra={'spider': spider}, *logformatter_adapter(logkws))
return self.signals.send_catch_log_deferred(signal=signals.item_dropped, item=item, response=response, spider=spider, exception=output.value)
else:
logger.error('Error processing %(item)s', {'item': item}, exc_info=failure_to_exc_info(output), extra={'spider': spider})
else:
logkws = self.logformatter.scraped(output, response, spider)
logger.log(extra={'spider': spider}, *logformatter_adapter(logkws))
return self.signals.send_catch_log_deferred(signal=signals.item_scraped, item=output, response=response, spider=spider)
|
'Lazy-load the downloadhandler for a scheme
only on the first request for that scheme.'
| def _get_handler(self, scheme):
| if (scheme in self._handlers):
return self._handlers[scheme]
if (scheme in self._notconfigured):
return None
if (scheme not in self._schemes):
self._notconfigured[scheme] = 'no handler available for that scheme'
return None
path = self._schemes[scheme]
try:
dhcls = load_object(path)
dh = dhcls(self._crawler.settings)
except NotConfigured as ex:
self._notconfigured[scheme] = str(ex)
return None
except Exception as ex:
logger.error('Loading "%(clspath)s" for scheme "%(scheme)s"', {'clspath': path, 'scheme': scheme}, exc_info=True, extra={'crawler': self._crawler})
self._notconfigured[scheme] = str(ex)
return None
else:
self._handlers[scheme] = dh
return self._handlers[scheme]
|
'Return a deferred for the HTTP download'
| def download_request(self, request, spider):
| agent = ScrapyAgent(contextFactory=self._contextFactory, pool=self._pool, maxsize=getattr(spider, 'download_maxsize', self._default_maxsize), warnsize=getattr(spider, 'download_warnsize', self._default_warnsize), fail_on_dataloss=self._fail_on_dataloss)
return agent.download_request(request)
|
'Asks the proxy to open a tunnel.'
| def requestTunnel(self, protocol):
| tunnelReq = tunnel_request_data(self._tunneledHost, self._tunneledPort, self._proxyAuthHeader)
protocol.transport.write(tunnelReq)
self._protocolDataReceived = protocol.dataReceived
protocol.dataReceived = self.processProxyResponse
self._protocol = protocol
return protocol
|
'Processes the response from the proxy. If the tunnel is successfully
created, notifies the client that we are ready to send requests. If not
raises a TunnelError.'
| def processProxyResponse(self, rcvd_bytes):
| self._connectBuffer += rcvd_bytes
if ('\r\n\r\n' not in self._connectBuffer):
return
self._protocol.dataReceived = self._protocolDataReceived
respm = TunnelingTCP4ClientEndpoint._responseMatcher.match(self._connectBuffer)
if (respm and (int(respm.group('status')) == 200)):
try:
sslOptions = self._contextFactory.creatorForNetloc(self._tunneledHost, self._tunneledPort)
except AttributeError:
sslOptions = self._contextFactory
self._protocol.transport.startTLS(sslOptions, self._protocolFactory)
self._tunnelReadyDeferred.callback(self._protocol)
else:
if respm:
extra = {'status': int(respm.group('status')), 'reason': respm.group('reason').strip()}
else:
extra = rcvd_bytes[:32]
self._tunnelReadyDeferred.errback(TunnelError(('Could not open CONNECT tunnel with proxy %s:%s [%r]' % (self._host, self._port, extra))))
|
'Propagates the errback to the appropriate deferred.'
| def connectFailed(self, reason):
| self._tunnelReadyDeferred.errback(reason)
|
'Return a deferred for the HTTP download'
| def download_request(self, request, spider):
| factory = self.HTTPClientFactory(request)
self._connect(factory)
return factory.deferred
|
'Upload file to S3 storage'
| def persist_file(self, path, buf, info, meta=None, headers=None):
| key_name = ('%s%s' % (self.prefix, path))
buf.seek(0)
if self.is_botocore:
extra = self._headers_to_botocore_kwargs(self.HEADERS)
if headers:
extra.update(self._headers_to_botocore_kwargs(headers))
return threads.deferToThread(self.s3_client.put_object, Bucket=self.bucket, Key=key_name, Body=buf, Metadata={k: str(v) for (k, v) in six.iteritems((meta or {}))}, ACL=self.POLICY, **extra)
else:
b = self._get_boto_bucket()
k = b.new_key(key_name)
if meta:
for (metakey, metavalue) in six.iteritems(meta):
k.set_metadata(metakey, str(metavalue))
h = self.HEADERS.copy()
if headers:
h.update(headers)
return threads.deferToThread(k.set_contents_from_string, buf.getvalue(), headers=h, policy=self.POLICY)
|
'Convert headers to botocore keyword agruments.'
| def _headers_to_botocore_kwargs(self, headers):
| mapping = CaselessDict({'Content-Type': 'ContentType', 'Cache-Control': 'CacheControl', 'Content-Disposition': 'ContentDisposition', 'Content-Encoding': 'ContentEncoding', 'Content-Language': 'ContentLanguage', 'Content-Length': 'ContentLength', 'Content-MD5': 'ContentMD5', 'Expires': 'Expires', 'X-Amz-Grant-Full-Control': 'GrantFullControl', 'X-Amz-Grant-Read': 'GrantRead', 'X-Amz-Grant-Read-ACP': 'GrantReadACP', 'X-Amz-Grant-Write-ACP': 'GrantWriteACP'})
extra = {}
for (key, value) in six.iteritems(headers):
try:
kwarg = mapping[key]
except KeyError:
raise TypeError(('Header "%s" is not supported by botocore' % key))
else:
extra[kwarg] = value
return extra
|
End of preview. Expand
in Dataset Viewer.
YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/datasets-cards)
HF version of Edinburgh-NLP's Code docstrings corpus
- Downloads last month
- 20