commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 10
2.94k
| new_contents
stringlengths 21
3.18k
| subject
stringlengths 16
444
| message
stringlengths 17
2.63k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 5
43k
| ndiff
stringlengths 51
3.32k
| instruction
stringlengths 16
444
| content
stringlengths 133
4.32k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|
e905334869af72025592de586b81650cb3468b8a | sentry/queue/client.py | sentry/queue/client.py | from kombu import BrokerConnection
from kombu.common import maybe_declare
from kombu.pools import producers
from sentry.conf import settings
from sentry.queue.queues import task_queues, task_exchange
class Broker(object):
def __init__(self, config):
self.connection = BrokerConnection(**config)
def delay(self, func, *args, **kwargs):
payload = {
"func": func,
"args": args,
"kwargs": kwargs,
}
with producers[self.connection].acquire(block=False) as producer:
for queue in task_queues:
maybe_declare(queue, producer.channel)
producer.publish(payload,
exchange=task_exchange,
serializer="pickle",
compression="bzip2",
queue='default',
routing_key='default',
)
broker = Broker(settings.QUEUE)
| from kombu import BrokerConnection
from kombu.common import maybe_declare
from kombu.pools import producers
from sentry.conf import settings
from sentry.queue.queues import task_queues, task_exchange
class Broker(object):
def __init__(self, config):
self.connection = BrokerConnection(**config)
with producers[self.connection].acquire(block=False) as producer:
for queue in task_queues:
maybe_declare(queue, producer.channel)
def delay(self, func, *args, **kwargs):
payload = {
"func": func,
"args": args,
"kwargs": kwargs,
}
with producers[self.connection].acquire(block=False) as producer:
producer.publish(payload,
exchange=task_exchange,
serializer="pickle",
compression="bzip2",
queue='default',
routing_key='default',
)
broker = Broker(settings.QUEUE)
| Declare queues when broker is instantiated | Declare queues when broker is instantiated
| Python | bsd-3-clause | imankulov/sentry,BuildingLink/sentry,zenefits/sentry,korealerts1/sentry,kevinastone/sentry,fotinakis/sentry,fuziontech/sentry,ngonzalvez/sentry,mvaled/sentry,Kronuz/django-sentry,ngonzalvez/sentry,looker/sentry,felixbuenemann/sentry,ngonzalvez/sentry,nicholasserra/sentry,camilonova/sentry,jokey2k/sentry,llonchj/sentry,fuziontech/sentry,llonchj/sentry,NickPresta/sentry,boneyao/sentry,SilentCircle/sentry,Kryz/sentry,JamesMura/sentry,SilentCircle/sentry,wujuguang/sentry,JTCunning/sentry,rdio/sentry,1tush/sentry,alexm92/sentry,imankulov/sentry,wujuguang/sentry,jokey2k/sentry,jean/sentry,chayapan/django-sentry,looker/sentry,beeftornado/sentry,chayapan/django-sentry,gg7/sentry,chayapan/django-sentry,JamesMura/sentry,1tush/sentry,zenefits/sentry,ewdurbin/sentry,NickPresta/sentry,alex/sentry,camilonova/sentry,kevinastone/sentry,pauloschilling/sentry,boneyao/sentry,korealerts1/sentry,rdio/sentry,daevaorn/sentry,drcapulet/sentry,TedaLIEz/sentry,Kronuz/django-sentry,wong2/sentry,felixbuenemann/sentry,1tush/sentry,hongliang5623/sentry,looker/sentry,BayanGroup/sentry,BuildingLink/sentry,BuildingLink/sentry,BayanGroup/sentry,kevinlondon/sentry,daevaorn/sentry,looker/sentry,llonchj/sentry,gencer/sentry,fotinakis/sentry,BuildingLink/sentry,korealerts1/sentry,felixbuenemann/sentry,boneyao/sentry,alex/sentry,fuziontech/sentry,nicholasserra/sentry,beeftornado/sentry,SilentCircle/sentry,ifduyue/sentry,gencer/sentry,jean/sentry,pauloschilling/sentry,camilonova/sentry,wong2/sentry,mvaled/sentry,JamesMura/sentry,ewdurbin/sentry,JackDanger/sentry,beeftornado/sentry,BuildingLink/sentry,argonemyth/sentry,zenefits/sentry,alex/sentry,gg7/sentry,jean/sentry,vperron/sentry,NickPresta/sentry,mvaled/sentry,JackDanger/sentry,songyi199111/sentry,vperron/sentry,daevaorn/sentry,ifduyue/sentry,kevinlondon/sentry,BayanGroup/sentry,JackDanger/sentry,kevinastone/sentry,Natim/sentry,zenefits/sentry,pauloschilling/sentry,Natim/sentry,argonemyth/sentry,alexm92/sentry,Natim/sentry,Kronuz/django-sentry,gg7/sentry,hongliang5623/sentry,looker/sentry,daevaorn/sentry,TedaLIEz/sentry,alexm92/sentry,vperron/sentry,ifduyue/sentry,mvaled/sentry,jean/sentry,JamesMura/sentry,hongliang5623/sentry,songyi199111/sentry,argonemyth/sentry,JTCunning/sentry,gencer/sentry,nicholasserra/sentry,ifduyue/sentry,Kryz/sentry,Kryz/sentry,beni55/sentry,TedaLIEz/sentry,kevinlondon/sentry,NickPresta/sentry,mvaled/sentry,drcapulet/sentry,jean/sentry,gencer/sentry,songyi199111/sentry,beni55/sentry,beni55/sentry,SilentCircle/sentry,mvaled/sentry,mitsuhiko/sentry,ewdurbin/sentry,wujuguang/sentry,fotinakis/sentry,drcapulet/sentry,fotinakis/sentry,wong2/sentry,ifduyue/sentry,zenefits/sentry,JTCunning/sentry,rdio/sentry,imankulov/sentry,jokey2k/sentry,gencer/sentry,rdio/sentry,JamesMura/sentry,mitsuhiko/sentry | from kombu import BrokerConnection
from kombu.common import maybe_declare
from kombu.pools import producers
from sentry.conf import settings
from sentry.queue.queues import task_queues, task_exchange
class Broker(object):
def __init__(self, config):
self.connection = BrokerConnection(**config)
+ with producers[self.connection].acquire(block=False) as producer:
+ for queue in task_queues:
+ maybe_declare(queue, producer.channel)
def delay(self, func, *args, **kwargs):
payload = {
"func": func,
"args": args,
"kwargs": kwargs,
}
with producers[self.connection].acquire(block=False) as producer:
- for queue in task_queues:
- maybe_declare(queue, producer.channel)
producer.publish(payload,
exchange=task_exchange,
serializer="pickle",
compression="bzip2",
queue='default',
routing_key='default',
)
broker = Broker(settings.QUEUE)
| Declare queues when broker is instantiated | ## Code Before:
from kombu import BrokerConnection
from kombu.common import maybe_declare
from kombu.pools import producers
from sentry.conf import settings
from sentry.queue.queues import task_queues, task_exchange
class Broker(object):
def __init__(self, config):
self.connection = BrokerConnection(**config)
def delay(self, func, *args, **kwargs):
payload = {
"func": func,
"args": args,
"kwargs": kwargs,
}
with producers[self.connection].acquire(block=False) as producer:
for queue in task_queues:
maybe_declare(queue, producer.channel)
producer.publish(payload,
exchange=task_exchange,
serializer="pickle",
compression="bzip2",
queue='default',
routing_key='default',
)
broker = Broker(settings.QUEUE)
## Instruction:
Declare queues when broker is instantiated
## Code After:
from kombu import BrokerConnection
from kombu.common import maybe_declare
from kombu.pools import producers
from sentry.conf import settings
from sentry.queue.queues import task_queues, task_exchange
class Broker(object):
def __init__(self, config):
self.connection = BrokerConnection(**config)
with producers[self.connection].acquire(block=False) as producer:
for queue in task_queues:
maybe_declare(queue, producer.channel)
def delay(self, func, *args, **kwargs):
payload = {
"func": func,
"args": args,
"kwargs": kwargs,
}
with producers[self.connection].acquire(block=False) as producer:
producer.publish(payload,
exchange=task_exchange,
serializer="pickle",
compression="bzip2",
queue='default',
routing_key='default',
)
broker = Broker(settings.QUEUE)
|
950ac9130bafe1fced578bf61d746b047830bfa0 | automata/base/exceptions.py | automata/base/exceptions.py | """Exception classes shared by all automata."""
class AutomatonException(Exception):
"""The base class for all automaton-related errors."""
pass
class InvalidStateError(AutomatonException):
"""A state is not a valid state for this automaton."""
pass
class InvalidSymbolError(AutomatonException):
"""A symbol is not a valid symbol for this automaton."""
pass
class MissingStateError(AutomatonException):
"""A state is missing from the automaton definition."""
pass
class MissingSymbolError(AutomatonException):
"""A symbol is missing from the automaton definition."""
pass
class InitialStateError(AutomatonException):
"""The initial state fails to meet some required condition."""
pass
class FinalStateError(AutomatonException):
"""A final state fails to meet some required condition."""
pass
class RejectionException(AutomatonException):
"""The input was rejected by the automaton after validation."""
pass
| """Exception classes shared by all automata."""
class AutomatonException(Exception):
"""The base class for all automaton-related errors."""
pass
class InvalidStateError(AutomatonException):
"""A state is not a valid state for this automaton."""
pass
class InvalidSymbolError(AutomatonException):
"""A symbol is not a valid symbol for this automaton."""
pass
class MissingStateError(AutomatonException):
"""A state is missing from the automaton definition."""
pass
class MissingSymbolError(AutomatonException):
"""A symbol is missing from the automaton definition."""
pass
class InitialStateError(AutomatonException):
"""The initial state fails to meet some required condition."""
pass
class FinalStateError(AutomatonException):
"""A final state fails to meet some required condition."""
pass
class RejectionException(AutomatonException):
"""The input was rejected by the automaton."""
pass
| Remove "validation" from RejectionException docstring | Remove "validation" from RejectionException docstring
| Python | mit | caleb531/automata | """Exception classes shared by all automata."""
class AutomatonException(Exception):
"""The base class for all automaton-related errors."""
pass
class InvalidStateError(AutomatonException):
"""A state is not a valid state for this automaton."""
pass
class InvalidSymbolError(AutomatonException):
"""A symbol is not a valid symbol for this automaton."""
pass
class MissingStateError(AutomatonException):
"""A state is missing from the automaton definition."""
pass
class MissingSymbolError(AutomatonException):
"""A symbol is missing from the automaton definition."""
pass
class InitialStateError(AutomatonException):
"""The initial state fails to meet some required condition."""
pass
class FinalStateError(AutomatonException):
"""A final state fails to meet some required condition."""
pass
class RejectionException(AutomatonException):
- """The input was rejected by the automaton after validation."""
+ """The input was rejected by the automaton."""
pass
| Remove "validation" from RejectionException docstring | ## Code Before:
"""Exception classes shared by all automata."""
class AutomatonException(Exception):
"""The base class for all automaton-related errors."""
pass
class InvalidStateError(AutomatonException):
"""A state is not a valid state for this automaton."""
pass
class InvalidSymbolError(AutomatonException):
"""A symbol is not a valid symbol for this automaton."""
pass
class MissingStateError(AutomatonException):
"""A state is missing from the automaton definition."""
pass
class MissingSymbolError(AutomatonException):
"""A symbol is missing from the automaton definition."""
pass
class InitialStateError(AutomatonException):
"""The initial state fails to meet some required condition."""
pass
class FinalStateError(AutomatonException):
"""A final state fails to meet some required condition."""
pass
class RejectionException(AutomatonException):
"""The input was rejected by the automaton after validation."""
pass
## Instruction:
Remove "validation" from RejectionException docstring
## Code After:
"""Exception classes shared by all automata."""
class AutomatonException(Exception):
"""The base class for all automaton-related errors."""
pass
class InvalidStateError(AutomatonException):
"""A state is not a valid state for this automaton."""
pass
class InvalidSymbolError(AutomatonException):
"""A symbol is not a valid symbol for this automaton."""
pass
class MissingStateError(AutomatonException):
"""A state is missing from the automaton definition."""
pass
class MissingSymbolError(AutomatonException):
"""A symbol is missing from the automaton definition."""
pass
class InitialStateError(AutomatonException):
"""The initial state fails to meet some required condition."""
pass
class FinalStateError(AutomatonException):
"""A final state fails to meet some required condition."""
pass
class RejectionException(AutomatonException):
"""The input was rejected by the automaton."""
pass
|
c3f8860c717a139d396b0d902db989ab7b8369ba | stock_inventory_hierarchical/__openerp__.py | stock_inventory_hierarchical/__openerp__.py |
{
"name": "Hierarchical Inventory adjustments",
"summary": "Group several Inventory adjustments in a master inventory",
"version": "8.0.2.0.0",
"depends": ["stock"],
"author": u"Numérigraphe,Odoo Community Association (OCA)",
"category": "Warehouse Management",
"data": ["views/stock_inventory_view.xml",
"wizard/generate_inventory_view.xml"],
"images": ["inventory_form.png",
"inventory_form_actions.png",
"wizard.png"],
'license': 'AGPL-3',
'installable': True
}
|
{
"name": "Hierarchical Inventory adjustments",
"summary": "Group several Inventory adjustments in a master inventory",
"version": "8.0.2.0.0",
"depends": ["stock"],
"author": u"Numérigraphe,Odoo Community Association (OCA)",
"category": "Warehouse Management",
"data": ["views/stock_inventory_view.xml",
"wizard/generate_inventory_view.xml"],
"images": ["images/inventory_form.png",
"images/inventory_form_actions.png",
"images/wizard.png"],
'license': 'AGPL-3',
'installable': True
}
| Fix image path in manifest | Fix image path in manifest
| Python | agpl-3.0 | kmee/stock-logistics-warehouse,factorlibre/stock-logistics-warehouse,open-synergy/stock-logistics-warehouse,acsone/stock-logistics-warehouse,avoinsystems/stock-logistics-warehouse |
{
"name": "Hierarchical Inventory adjustments",
"summary": "Group several Inventory adjustments in a master inventory",
"version": "8.0.2.0.0",
"depends": ["stock"],
"author": u"Numérigraphe,Odoo Community Association (OCA)",
"category": "Warehouse Management",
"data": ["views/stock_inventory_view.xml",
"wizard/generate_inventory_view.xml"],
- "images": ["inventory_form.png",
+ "images": ["images/inventory_form.png",
- "inventory_form_actions.png",
+ "images/inventory_form_actions.png",
- "wizard.png"],
+ "images/wizard.png"],
'license': 'AGPL-3',
'installable': True
}
| Fix image path in manifest | ## Code Before:
{
"name": "Hierarchical Inventory adjustments",
"summary": "Group several Inventory adjustments in a master inventory",
"version": "8.0.2.0.0",
"depends": ["stock"],
"author": u"Numérigraphe,Odoo Community Association (OCA)",
"category": "Warehouse Management",
"data": ["views/stock_inventory_view.xml",
"wizard/generate_inventory_view.xml"],
"images": ["inventory_form.png",
"inventory_form_actions.png",
"wizard.png"],
'license': 'AGPL-3',
'installable': True
}
## Instruction:
Fix image path in manifest
## Code After:
{
"name": "Hierarchical Inventory adjustments",
"summary": "Group several Inventory adjustments in a master inventory",
"version": "8.0.2.0.0",
"depends": ["stock"],
"author": u"Numérigraphe,Odoo Community Association (OCA)",
"category": "Warehouse Management",
"data": ["views/stock_inventory_view.xml",
"wizard/generate_inventory_view.xml"],
"images": ["images/inventory_form.png",
"images/inventory_form_actions.png",
"images/wizard.png"],
'license': 'AGPL-3',
'installable': True
}
|
70a642c0597fb2f929fc83d821c8b1f095ed1328 | proxy/plugins/plugins.py | proxy/plugins/plugins.py | packetFunctions = {}
commands = {}
onStart = []
onConnection = []
onConnectionLoss = []
class packetHook(object):
def __init__(self, pktType, pktSubtype):
self.pktType = pktType
self.pktSubtype = pktSubtype
def __call__(self, f):
global packetFunctions
packetFunctions[(self.pktType, self.pktSubtype)] = f
class commandHook(object):
"""docstring for commandHook"""
def __init__(self, command):
self.command = command
def __call__(self, f):
global commands
commands[self.command] = f
def onStartHook(f):
global onStart
onStart.append(f)
return f
def onConnectionHook(f):
global onConnection
onConnection.append(f)
return f
def onConnectionLossHook(f):
global onConnectionLoss
onConnectionLoss.append(f)
return f | packetFunctions = {}
commands = {}
onStart = []
onConnection = []
onConnectionLoss = []
class packetHook(object):
def __init__(self, pktType, pktSubtype):
self.pktType = pktType
self.pktSubtype = pktSubtype
def __call__(self, f):
global packetFunctions
if (self.pktType, self.pktSubtype) not in packetFunctions:
packetFunctions[(self.pktType, self.pktSubtype)] = []
packetFunctions[(self.pktType, self.pktSubtype)].append(f)
class commandHook(object):
"""docstring for commandHook"""
def __init__(self, command):
self.command = command
def __call__(self, f):
global commands
commands[self.command] = f
def onStartHook(f):
global onStart
onStart.append(f)
return f
def onConnectionHook(f):
global onConnection
onConnection.append(f)
return f
def onConnectionLossHook(f):
global onConnectionLoss
onConnectionLoss.append(f)
return f | Allow mutiple hooks for packets | Allow mutiple hooks for packets
| Python | agpl-3.0 | alama/PSO2Proxy,alama/PSO2Proxy,flyergo/PSO2Proxy,alama/PSO2Proxy,cyberkitsune/PSO2Proxy,cyberkitsune/PSO2Proxy,flyergo/PSO2Proxy,cyberkitsune/PSO2Proxy | packetFunctions = {}
commands = {}
onStart = []
onConnection = []
onConnectionLoss = []
class packetHook(object):
def __init__(self, pktType, pktSubtype):
self.pktType = pktType
self.pktSubtype = pktSubtype
def __call__(self, f):
global packetFunctions
+ if (self.pktType, self.pktSubtype) not in packetFunctions:
- packetFunctions[(self.pktType, self.pktSubtype)] = f
+ packetFunctions[(self.pktType, self.pktSubtype)] = []
+ packetFunctions[(self.pktType, self.pktSubtype)].append(f)
class commandHook(object):
"""docstring for commandHook"""
def __init__(self, command):
self.command = command
def __call__(self, f):
global commands
commands[self.command] = f
def onStartHook(f):
global onStart
onStart.append(f)
return f
def onConnectionHook(f):
global onConnection
onConnection.append(f)
return f
def onConnectionLossHook(f):
global onConnectionLoss
onConnectionLoss.append(f)
return f | Allow mutiple hooks for packets | ## Code Before:
packetFunctions = {}
commands = {}
onStart = []
onConnection = []
onConnectionLoss = []
class packetHook(object):
def __init__(self, pktType, pktSubtype):
self.pktType = pktType
self.pktSubtype = pktSubtype
def __call__(self, f):
global packetFunctions
packetFunctions[(self.pktType, self.pktSubtype)] = f
class commandHook(object):
"""docstring for commandHook"""
def __init__(self, command):
self.command = command
def __call__(self, f):
global commands
commands[self.command] = f
def onStartHook(f):
global onStart
onStart.append(f)
return f
def onConnectionHook(f):
global onConnection
onConnection.append(f)
return f
def onConnectionLossHook(f):
global onConnectionLoss
onConnectionLoss.append(f)
return f
## Instruction:
Allow mutiple hooks for packets
## Code After:
packetFunctions = {}
commands = {}
onStart = []
onConnection = []
onConnectionLoss = []
class packetHook(object):
def __init__(self, pktType, pktSubtype):
self.pktType = pktType
self.pktSubtype = pktSubtype
def __call__(self, f):
global packetFunctions
if (self.pktType, self.pktSubtype) not in packetFunctions:
packetFunctions[(self.pktType, self.pktSubtype)] = []
packetFunctions[(self.pktType, self.pktSubtype)].append(f)
class commandHook(object):
"""docstring for commandHook"""
def __init__(self, command):
self.command = command
def __call__(self, f):
global commands
commands[self.command] = f
def onStartHook(f):
global onStart
onStart.append(f)
return f
def onConnectionHook(f):
global onConnection
onConnection.append(f)
return f
def onConnectionLossHook(f):
global onConnectionLoss
onConnectionLoss.append(f)
return f |
af91b7c2612fab598ba50c0c0256f7e552098d92 | reportlab/docs/genAll.py | reportlab/docs/genAll.py | """Runs the three manual-building scripts"""
if __name__=='__main__':
import os, sys
d = os.path.dirname(sys.argv[0])
#need a quiet mode for the test suite
if '-s' in sys.argv: # 'silent
quiet = '-s'
else:
quiet = ''
if not d: d = '.'
if not os.path.isabs(d):
d = os.path.normpath(os.path.join(os.getcwd(),d))
for p in ('reference/genreference.py',
'userguide/genuserguide.py',
'graphguide/gengraphguide.py',
'../tools/docco/graphdocpy.py'):
os.chdir(d)
os.chdir(os.path.dirname(p))
os.system('%s %s %s' % (sys.executable,os.path.basename(p), quiet))
| import os
def _genAll(d=None,quiet=''):
if not d: d = '.'
if not os.path.isabs(d):
d = os.path.normpath(os.path.join(os.getcwd(),d))
for p in ('reference/genreference.py',
'userguide/genuserguide.py',
'graphguide/gengraphguide.py',
'../tools/docco/graphdocpy.py'):
os.chdir(d)
os.chdir(os.path.dirname(p))
os.system('%s %s %s' % (sys.executable,os.path.basename(p), quiet))
"""Runs the manual-building scripts"""
if __name__=='__main__':
import sys
#need a quiet mode for the test suite
if '-s' in sys.argv: # 'silent
quiet = '-s'
else:
quiet = ''
_genAll(os.path.dirname(sys.argv[0]),quiet)
| Allow for use in daily.py | Allow for use in daily.py
| Python | bsd-3-clause | makinacorpus/reportlab-ecomobile,makinacorpus/reportlab-ecomobile,makinacorpus/reportlab-ecomobile,makinacorpus/reportlab-ecomobile,makinacorpus/reportlab-ecomobile | + import os
+ def _genAll(d=None,quiet=''):
+ if not d: d = '.'
+ if not os.path.isabs(d):
+ d = os.path.normpath(os.path.join(os.getcwd(),d))
+ for p in ('reference/genreference.py',
+ 'userguide/genuserguide.py',
+ 'graphguide/gengraphguide.py',
+ '../tools/docco/graphdocpy.py'):
+ os.chdir(d)
+ os.chdir(os.path.dirname(p))
+ os.system('%s %s %s' % (sys.executable,os.path.basename(p), quiet))
+
- """Runs the three manual-building scripts"""
+ """Runs the manual-building scripts"""
if __name__=='__main__':
- import os, sys
- d = os.path.dirname(sys.argv[0])
+ import sys
+ #need a quiet mode for the test suite
+ if '-s' in sys.argv: # 'silent
+ quiet = '-s'
+ else:
+ quiet = ''
+ _genAll(os.path.dirname(sys.argv[0]),quiet)
- #need a quiet mode for the test suite
- if '-s' in sys.argv: # 'silent
- quiet = '-s'
- else:
- quiet = ''
-
- if not d: d = '.'
- if not os.path.isabs(d):
- d = os.path.normpath(os.path.join(os.getcwd(),d))
- for p in ('reference/genreference.py',
- 'userguide/genuserguide.py',
- 'graphguide/gengraphguide.py',
- '../tools/docco/graphdocpy.py'):
- os.chdir(d)
- os.chdir(os.path.dirname(p))
- os.system('%s %s %s' % (sys.executable,os.path.basename(p), quiet))
- | Allow for use in daily.py | ## Code Before:
"""Runs the three manual-building scripts"""
if __name__=='__main__':
import os, sys
d = os.path.dirname(sys.argv[0])
#need a quiet mode for the test suite
if '-s' in sys.argv: # 'silent
quiet = '-s'
else:
quiet = ''
if not d: d = '.'
if not os.path.isabs(d):
d = os.path.normpath(os.path.join(os.getcwd(),d))
for p in ('reference/genreference.py',
'userguide/genuserguide.py',
'graphguide/gengraphguide.py',
'../tools/docco/graphdocpy.py'):
os.chdir(d)
os.chdir(os.path.dirname(p))
os.system('%s %s %s' % (sys.executable,os.path.basename(p), quiet))
## Instruction:
Allow for use in daily.py
## Code After:
import os
def _genAll(d=None,quiet=''):
if not d: d = '.'
if not os.path.isabs(d):
d = os.path.normpath(os.path.join(os.getcwd(),d))
for p in ('reference/genreference.py',
'userguide/genuserguide.py',
'graphguide/gengraphguide.py',
'../tools/docco/graphdocpy.py'):
os.chdir(d)
os.chdir(os.path.dirname(p))
os.system('%s %s %s' % (sys.executable,os.path.basename(p), quiet))
"""Runs the manual-building scripts"""
if __name__=='__main__':
import sys
#need a quiet mode for the test suite
if '-s' in sys.argv: # 'silent
quiet = '-s'
else:
quiet = ''
_genAll(os.path.dirname(sys.argv[0]),quiet)
|
e34969db596ff3dfa4bf78efb3f3ccfe771d9ef9 | setup.py | setup.py | try:
from setuptools.core import setup
except ImportError:
from distutils.core import setup
PACKAGE = 'django_exceptional_middleware'
VERSION = '0.2'
data_files = [
(
'exceptional_middleware/templates/http_responses', [ 'exceptional_middleware/templates/http_responses/default.html' ],
),
]
setup(
name=PACKAGE, version=VERSION,
description="Django middleware to allow generating arbitrary HTTP status codes via exceptions.",
packages=[ 'exceptional_middleware' ],
data_files=data_files,
license='MIT',
author='James Aylett',
url = 'http://tartarus.org/james/computers/django/',
)
| try:
from setuptools.core import setup
except ImportError:
from distutils.core import setup
PACKAGE = 'django_exceptional_middleware'
VERSION = '0.4'
package_data = {
'exceptional_middleware': [ 'templates/http_responses/*.html' ],
}
setup(
name=PACKAGE, version=VERSION,
description="Django middleware to allow generating arbitrary HTTP status codes via exceptions.",
packages=[ 'exceptional_middleware' ],
package_data=package_data,
license='MIT',
author='James Aylett',
url = 'http://tartarus.org/james/computers/django/',
)
| Fix templates install. Bump to version 0.4 in the process (which is really my laziness). | Fix templates install. Bump to version 0.4 in the process (which is really my laziness).
| Python | mit | jaylett/django_exceptional_middleware | try:
from setuptools.core import setup
except ImportError:
from distutils.core import setup
PACKAGE = 'django_exceptional_middleware'
- VERSION = '0.2'
+ VERSION = '0.4'
+ package_data = {
+ 'exceptional_middleware': [ 'templates/http_responses/*.html' ],
+ }
- data_files = [
- (
- 'exceptional_middleware/templates/http_responses', [ 'exceptional_middleware/templates/http_responses/default.html' ],
- ),
- ]
setup(
name=PACKAGE, version=VERSION,
description="Django middleware to allow generating arbitrary HTTP status codes via exceptions.",
packages=[ 'exceptional_middleware' ],
- data_files=data_files,
+ package_data=package_data,
license='MIT',
author='James Aylett',
url = 'http://tartarus.org/james/computers/django/',
)
| Fix templates install. Bump to version 0.4 in the process (which is really my laziness). | ## Code Before:
try:
from setuptools.core import setup
except ImportError:
from distutils.core import setup
PACKAGE = 'django_exceptional_middleware'
VERSION = '0.2'
data_files = [
(
'exceptional_middleware/templates/http_responses', [ 'exceptional_middleware/templates/http_responses/default.html' ],
),
]
setup(
name=PACKAGE, version=VERSION,
description="Django middleware to allow generating arbitrary HTTP status codes via exceptions.",
packages=[ 'exceptional_middleware' ],
data_files=data_files,
license='MIT',
author='James Aylett',
url = 'http://tartarus.org/james/computers/django/',
)
## Instruction:
Fix templates install. Bump to version 0.4 in the process (which is really my laziness).
## Code After:
try:
from setuptools.core import setup
except ImportError:
from distutils.core import setup
PACKAGE = 'django_exceptional_middleware'
VERSION = '0.4'
package_data = {
'exceptional_middleware': [ 'templates/http_responses/*.html' ],
}
setup(
name=PACKAGE, version=VERSION,
description="Django middleware to allow generating arbitrary HTTP status codes via exceptions.",
packages=[ 'exceptional_middleware' ],
package_data=package_data,
license='MIT',
author='James Aylett',
url = 'http://tartarus.org/james/computers/django/',
)
|
c6265c2112ee9985af8b6b80fe0bee1811dc6abd | setup.py | setup.py | from distutils.core import setup
setup(
name='oceanoptics',
version='0.2.6',
author='Andreas Poehlmann, Jose A. Jimenez-Berni, Ben Gamari, Simon Dickreuter',
author_email='mail@andreaspoehlmann.de',
packages=['oceanoptics', 'oceanoptics.spectrometers'],
description='A Python driver for Ocean Optics spectrometers.',
long_description=open('README.md').read(),
requires=['python (>= 2.7)', 'pyusb (>= 1.0)', 'numpy'],
)
| from distutils.core import setup
setup(
name='oceanoptics',
version='0.2.7',
author='Andreas Poehlmann, Jose A. Jimenez-Berni, Ben Gamari, Simon Dickreuter, Ian Ross Williams',
author_email='mail@andreaspoehlmann.de',
packages=['oceanoptics', 'oceanoptics.spectrometers'],
description='A Python driver for Ocean Optics spectrometers.',
long_description=open('README.md').read(),
requires=['python (>= 2.7)', 'pyusb (>= 1.0)', 'numpy'],
)
| Add author and increase version number | Add author and increase version number
| Python | mit | ap--/python-oceanoptics | from distutils.core import setup
setup(
name='oceanoptics',
- version='0.2.6',
+ version='0.2.7',
- author='Andreas Poehlmann, Jose A. Jimenez-Berni, Ben Gamari, Simon Dickreuter',
+ author='Andreas Poehlmann, Jose A. Jimenez-Berni, Ben Gamari, Simon Dickreuter, Ian Ross Williams',
author_email='mail@andreaspoehlmann.de',
packages=['oceanoptics', 'oceanoptics.spectrometers'],
description='A Python driver for Ocean Optics spectrometers.',
long_description=open('README.md').read(),
requires=['python (>= 2.7)', 'pyusb (>= 1.0)', 'numpy'],
)
| Add author and increase version number | ## Code Before:
from distutils.core import setup
setup(
name='oceanoptics',
version='0.2.6',
author='Andreas Poehlmann, Jose A. Jimenez-Berni, Ben Gamari, Simon Dickreuter',
author_email='mail@andreaspoehlmann.de',
packages=['oceanoptics', 'oceanoptics.spectrometers'],
description='A Python driver for Ocean Optics spectrometers.',
long_description=open('README.md').read(),
requires=['python (>= 2.7)', 'pyusb (>= 1.0)', 'numpy'],
)
## Instruction:
Add author and increase version number
## Code After:
from distutils.core import setup
setup(
name='oceanoptics',
version='0.2.7',
author='Andreas Poehlmann, Jose A. Jimenez-Berni, Ben Gamari, Simon Dickreuter, Ian Ross Williams',
author_email='mail@andreaspoehlmann.de',
packages=['oceanoptics', 'oceanoptics.spectrometers'],
description='A Python driver for Ocean Optics spectrometers.',
long_description=open('README.md').read(),
requires=['python (>= 2.7)', 'pyusb (>= 1.0)', 'numpy'],
)
|
5bb9b2c9d5df410c85f4736c17224aeb2f05dd33 | s2v3.py | s2v3.py | from s2v2 import *
def calculate_sum(data_sample):
total = 0
for row in data_sample[1:]: # slice to start at row two, but I think we should only skip row 1 if we're importing the full csv (data_from_csv), but if we use the data w/ the header (my_csv) we'll be skipping a row that we're not supposed to skip (the actual first row of non-header data).
price = float(row[2])
total += price
return total
print('the sum total of prices for all ties in the dataset = ' + str(calculate_sum(data_from_csv))) # ok we're using the right import, but having two imports is confusing. | from s2v2 import *
def calculate_sum(data_sample):
total = 0
for row in data_sample[1:]: # slice to start at row two, but I think we should only skip row 1 if we're importing the full csv (data_from_csv), but if we use the data w/ the header (my_csv) we'll be skipping a row that we're not supposed to skip (the actual first row of non-header data).
price = float(row[2])
total += price
return total
print('the sum total of prices for all ties in the dataset = ', calculate_sum(data_from_csv)) # ok we're using the right import, but having two imports is confusing. UPDDATE: No, I don't have to convert the calculate_sum result to a string to add text about it, I just need to use , instead of + | Update print result to use "," instead of "+" for context text | Update print result to use "," instead of "+" for context text
| Python | mit | alexmilesyounger/ds_basics | from s2v2 import *
def calculate_sum(data_sample):
total = 0
for row in data_sample[1:]: # slice to start at row two, but I think we should only skip row 1 if we're importing the full csv (data_from_csv), but if we use the data w/ the header (my_csv) we'll be skipping a row that we're not supposed to skip (the actual first row of non-header data).
price = float(row[2])
total += price
return total
- print('the sum total of prices for all ties in the dataset = ' + str(calculate_sum(data_from_csv))) # ok we're using the right import, but having two imports is confusing.
+ print('the sum total of prices for all ties in the dataset = ', calculate_sum(data_from_csv)) # ok we're using the right import, but having two imports is confusing. UPDDATE: No, I don't have to convert the calculate_sum result to a string to add text about it, I just need to use , instead of + | Update print result to use "," instead of "+" for context text | ## Code Before:
from s2v2 import *
def calculate_sum(data_sample):
total = 0
for row in data_sample[1:]: # slice to start at row two, but I think we should only skip row 1 if we're importing the full csv (data_from_csv), but if we use the data w/ the header (my_csv) we'll be skipping a row that we're not supposed to skip (the actual first row of non-header data).
price = float(row[2])
total += price
return total
print('the sum total of prices for all ties in the dataset = ' + str(calculate_sum(data_from_csv))) # ok we're using the right import, but having two imports is confusing.
## Instruction:
Update print result to use "," instead of "+" for context text
## Code After:
from s2v2 import *
def calculate_sum(data_sample):
total = 0
for row in data_sample[1:]: # slice to start at row two, but I think we should only skip row 1 if we're importing the full csv (data_from_csv), but if we use the data w/ the header (my_csv) we'll be skipping a row that we're not supposed to skip (the actual first row of non-header data).
price = float(row[2])
total += price
return total
print('the sum total of prices for all ties in the dataset = ', calculate_sum(data_from_csv)) # ok we're using the right import, but having two imports is confusing. UPDDATE: No, I don't have to convert the calculate_sum result to a string to add text about it, I just need to use , instead of + |
3ba5b6491bf61e2d2919f05bbf5cef088a754aeb | molecule/default/tests/test_installation.py | molecule/default/tests/test_installation.py |
import os
import pytest
from testinfra.utils.ansible_runner import AnsibleRunner
testinfra_hosts = AnsibleRunner(
os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all')
@pytest.mark.parametrize('name', [
('vsftpd'),
('db5.3-util'),
])
def test_installed_packages(host, name):
"""
Test if packages installed
"""
assert host.package(name).is_installed
def test_service(host):
"""
Test service state
"""
service = host.service('vsftpd')
assert service.is_enabled
# if host.system_info.codename in ['jessie', 'xenial']:
if host.file('/etc/init.d/vsftpd').exists:
assert 'is running' in host.check_output('/etc/init.d/vsftpd status')
else:
assert service.is_running
def test_process(host):
"""
Test process state
"""
assert len(host.process.filter(comm='vsftpd')) == 1
def test_socket(host):
"""
Test ports
"""
assert host.socket('tcp://127.0.0.1:21').is_listening
def test_user(host):
"""
Test ftp user exists
"""
ftp_user = host.user('ftp')
assert ftp_user.exists
assert ftp_user.shell == '/bin/false'
|
import os
import pytest
from testinfra.utils.ansible_runner import AnsibleRunner
testinfra_hosts = AnsibleRunner(
os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all')
@pytest.mark.parametrize('name', [
('vsftpd'),
('db5.3-util'),
])
def test_installed_packages(host, name):
"""
Test if packages installed
"""
assert host.package(name).is_installed
def test_service(host):
"""
Test service state
"""
service = host.service('vsftpd')
assert service.is_enabled
# if host.system_info.codename in ['jessie', 'xenial']:
if host.file('/etc/init.d/vsftpd').exists:
assert 'is running' in host.check_output('/etc/init.d/vsftpd status')
else:
assert service.is_running
def test_process(host):
"""
Test process state
"""
assert len(host.process.filter(comm='vsftpd')) == 1
def test_socket(host):
"""
Test ports
"""
assert host.socket('tcp://127.0.0.1:21').is_listening
def test_user(host):
"""
Test ftp user exists
"""
ftp_user = host.user('ftp')
assert ftp_user.exists
assert ftp_user.shell in ['/usr/sbin/nologin', '/bin/false']
| Add nologin in expected user shell test | Add nologin in expected user shell test
| Python | mit | infOpen/ansible-role-vsftpd |
import os
import pytest
from testinfra.utils.ansible_runner import AnsibleRunner
testinfra_hosts = AnsibleRunner(
os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all')
@pytest.mark.parametrize('name', [
('vsftpd'),
('db5.3-util'),
])
def test_installed_packages(host, name):
"""
Test if packages installed
"""
assert host.package(name).is_installed
def test_service(host):
"""
Test service state
"""
service = host.service('vsftpd')
assert service.is_enabled
# if host.system_info.codename in ['jessie', 'xenial']:
if host.file('/etc/init.d/vsftpd').exists:
assert 'is running' in host.check_output('/etc/init.d/vsftpd status')
else:
assert service.is_running
def test_process(host):
"""
Test process state
"""
assert len(host.process.filter(comm='vsftpd')) == 1
def test_socket(host):
"""
Test ports
"""
assert host.socket('tcp://127.0.0.1:21').is_listening
def test_user(host):
"""
Test ftp user exists
"""
ftp_user = host.user('ftp')
assert ftp_user.exists
- assert ftp_user.shell == '/bin/false'
+ assert ftp_user.shell in ['/usr/sbin/nologin', '/bin/false']
| Add nologin in expected user shell test | ## Code Before:
import os
import pytest
from testinfra.utils.ansible_runner import AnsibleRunner
testinfra_hosts = AnsibleRunner(
os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all')
@pytest.mark.parametrize('name', [
('vsftpd'),
('db5.3-util'),
])
def test_installed_packages(host, name):
"""
Test if packages installed
"""
assert host.package(name).is_installed
def test_service(host):
"""
Test service state
"""
service = host.service('vsftpd')
assert service.is_enabled
# if host.system_info.codename in ['jessie', 'xenial']:
if host.file('/etc/init.d/vsftpd').exists:
assert 'is running' in host.check_output('/etc/init.d/vsftpd status')
else:
assert service.is_running
def test_process(host):
"""
Test process state
"""
assert len(host.process.filter(comm='vsftpd')) == 1
def test_socket(host):
"""
Test ports
"""
assert host.socket('tcp://127.0.0.1:21').is_listening
def test_user(host):
"""
Test ftp user exists
"""
ftp_user = host.user('ftp')
assert ftp_user.exists
assert ftp_user.shell == '/bin/false'
## Instruction:
Add nologin in expected user shell test
## Code After:
import os
import pytest
from testinfra.utils.ansible_runner import AnsibleRunner
testinfra_hosts = AnsibleRunner(
os.environ['MOLECULE_INVENTORY_FILE']).get_hosts('all')
@pytest.mark.parametrize('name', [
('vsftpd'),
('db5.3-util'),
])
def test_installed_packages(host, name):
"""
Test if packages installed
"""
assert host.package(name).is_installed
def test_service(host):
"""
Test service state
"""
service = host.service('vsftpd')
assert service.is_enabled
# if host.system_info.codename in ['jessie', 'xenial']:
if host.file('/etc/init.d/vsftpd').exists:
assert 'is running' in host.check_output('/etc/init.d/vsftpd status')
else:
assert service.is_running
def test_process(host):
"""
Test process state
"""
assert len(host.process.filter(comm='vsftpd')) == 1
def test_socket(host):
"""
Test ports
"""
assert host.socket('tcp://127.0.0.1:21').is_listening
def test_user(host):
"""
Test ftp user exists
"""
ftp_user = host.user('ftp')
assert ftp_user.exists
assert ftp_user.shell in ['/usr/sbin/nologin', '/bin/false']
|
b4c9b76d132668695b77d37d7db3071e629fcba7 | makerscience_admin/models.py | makerscience_admin/models.py | from django.db import models
from solo.models import SingletonModel
class MakerScienceStaticContent (SingletonModel):
about = models.TextField(null=True, blank=True)
about_team = models.TextField(null=True, blank=True)
about_contact = models.TextField(null=True, blank=True)
about_faq = models.TextField(null=True, blank=True)
about_cgu = models.TextField(null=True, blank=True)
class PageViews(models.Model):
client = models.CharField(max_length=255)
resource_uri = models.CharField(max_length=255)
| from django.db import models
from django.db.models.signals import post_delete
from solo.models import SingletonModel
from accounts.models import ObjectProfileLink
from makerscience_forum.models import MakerSciencePost
class MakerScienceStaticContent (SingletonModel):
about = models.TextField(null=True, blank=True)
about_team = models.TextField(null=True, blank=True)
about_contact = models.TextField(null=True, blank=True)
about_faq = models.TextField(null=True, blank=True)
about_cgu = models.TextField(null=True, blank=True)
class PageViews(models.Model):
client = models.CharField(max_length=255)
resource_uri = models.CharField(max_length=255)
def clear_makerscience(sender, instance, **kwargs):
if sender == MakerSciencePost:
ObjectProfileLink.objects.filter(content_type__model='post',
object_id=instance.parent.id).delete()
instance.parent.delete()
post_delete.connect(clear_makerscience, sender=MakerSciencePost)
| Allow to clear useless instances | Allow to clear useless instances
| Python | agpl-3.0 | atiberghien/makerscience-server,atiberghien/makerscience-server | from django.db import models
+ from django.db.models.signals import post_delete
+
from solo.models import SingletonModel
+
+ from accounts.models import ObjectProfileLink
+
+ from makerscience_forum.models import MakerSciencePost
class MakerScienceStaticContent (SingletonModel):
about = models.TextField(null=True, blank=True)
about_team = models.TextField(null=True, blank=True)
about_contact = models.TextField(null=True, blank=True)
about_faq = models.TextField(null=True, blank=True)
about_cgu = models.TextField(null=True, blank=True)
class PageViews(models.Model):
client = models.CharField(max_length=255)
resource_uri = models.CharField(max_length=255)
+
+ def clear_makerscience(sender, instance, **kwargs):
+ if sender == MakerSciencePost:
+ ObjectProfileLink.objects.filter(content_type__model='post',
+ object_id=instance.parent.id).delete()
+ instance.parent.delete()
+
+ post_delete.connect(clear_makerscience, sender=MakerSciencePost)
+ | Allow to clear useless instances | ## Code Before:
from django.db import models
from solo.models import SingletonModel
class MakerScienceStaticContent (SingletonModel):
about = models.TextField(null=True, blank=True)
about_team = models.TextField(null=True, blank=True)
about_contact = models.TextField(null=True, blank=True)
about_faq = models.TextField(null=True, blank=True)
about_cgu = models.TextField(null=True, blank=True)
class PageViews(models.Model):
client = models.CharField(max_length=255)
resource_uri = models.CharField(max_length=255)
## Instruction:
Allow to clear useless instances
## Code After:
from django.db import models
from django.db.models.signals import post_delete
from solo.models import SingletonModel
from accounts.models import ObjectProfileLink
from makerscience_forum.models import MakerSciencePost
class MakerScienceStaticContent (SingletonModel):
about = models.TextField(null=True, blank=True)
about_team = models.TextField(null=True, blank=True)
about_contact = models.TextField(null=True, blank=True)
about_faq = models.TextField(null=True, blank=True)
about_cgu = models.TextField(null=True, blank=True)
class PageViews(models.Model):
client = models.CharField(max_length=255)
resource_uri = models.CharField(max_length=255)
def clear_makerscience(sender, instance, **kwargs):
if sender == MakerSciencePost:
ObjectProfileLink.objects.filter(content_type__model='post',
object_id=instance.parent.id).delete()
instance.parent.delete()
post_delete.connect(clear_makerscience, sender=MakerSciencePost)
|
c713273fe145418113d750579f8b135dc513c3b8 | config.py | config.py | import os
if os.environ.get('DATABASE_URL') is None:
SQLALCHEMY_DATABASE_URI = 'sqlite:///meetup.db'
else:
SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
| import os
SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
| Delete default case for SQLALCHEMY_DATABASE_URI | Delete default case for SQLALCHEMY_DATABASE_URI
if user doesn't set it, he coud have some problems with SQLite
| Python | mit | Stark-Mountain/meetup-facebook-bot,Stark-Mountain/meetup-facebook-bot | import os
- if os.environ.get('DATABASE_URL') is None:
- SQLALCHEMY_DATABASE_URI = 'sqlite:///meetup.db'
- else:
- SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
+ SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
-
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
| Delete default case for SQLALCHEMY_DATABASE_URI | ## Code Before:
import os
if os.environ.get('DATABASE_URL') is None:
SQLALCHEMY_DATABASE_URI = 'sqlite:///meetup.db'
else:
SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
## Instruction:
Delete default case for SQLALCHEMY_DATABASE_URI
## Code After:
import os
SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
|
c61d5e84863dd67b5b76ec8031e624642f4c957c | main.py | main.py | from .ide.command import plugin_unloaded
from .ide.error import *
from .ide.rebuild import *
from .ide.server import *
from .ide.settings import plugin_loaded
from .ide.text_command import *
from .ide.type_hints import *
from .ide.utility import *
| from .ide.auto_complete import *
from .ide.command import plugin_unloaded
from .ide.error import *
from .ide.rebuild import *
from .ide.server import *
from .ide.settings import plugin_loaded
from .ide.text_command import *
from .ide.type_hints import *
| Fix issue with wrong import | Fix issue with wrong import
| Python | mit | b123400/purescript-ide-sublime | + from .ide.auto_complete import *
from .ide.command import plugin_unloaded
from .ide.error import *
from .ide.rebuild import *
from .ide.server import *
from .ide.settings import plugin_loaded
from .ide.text_command import *
from .ide.type_hints import *
- from .ide.utility import *
| Fix issue with wrong import | ## Code Before:
from .ide.command import plugin_unloaded
from .ide.error import *
from .ide.rebuild import *
from .ide.server import *
from .ide.settings import plugin_loaded
from .ide.text_command import *
from .ide.type_hints import *
from .ide.utility import *
## Instruction:
Fix issue with wrong import
## Code After:
from .ide.auto_complete import *
from .ide.command import plugin_unloaded
from .ide.error import *
from .ide.rebuild import *
from .ide.server import *
from .ide.settings import plugin_loaded
from .ide.text_command import *
from .ide.type_hints import *
|
4e75e742475236cf7358b4481a29a54eb607dd4d | spacy/tests/regression/test_issue850.py | spacy/tests/regression/test_issue850.py | '''
Test Matcher matches with '*' operator and Boolean flag
'''
from __future__ import unicode_literals
import pytest
from ...matcher import Matcher
from ...vocab import Vocab
from ...attrs import LOWER
from ...tokens import Doc
@pytest.mark.xfail
def test_issue850():
matcher = Matcher(Vocab())
IS_ANY_TOKEN = matcher.vocab.add_flag(lambda x: True)
matcher.add_pattern(
"FarAway",
[
{LOWER: "bob"},
{'OP': '*', IS_ANY_TOKEN: True},
{LOWER: 'frank'}
])
doc = Doc(matcher.vocab, words=['bob', 'and', 'and', 'cat', 'frank'])
match = matcher(doc)
assert len(match) == 1
start, end, label, ent_id = match
assert start == 0
assert end == 4
| '''
Test Matcher matches with '*' operator and Boolean flag
'''
from __future__ import unicode_literals
from __future__ import print_function
import pytest
from ...matcher import Matcher
from ...vocab import Vocab
from ...attrs import LOWER
from ...tokens import Doc
def test_basic_case():
matcher = Matcher(Vocab(
lex_attr_getters={LOWER: lambda string: string.lower()}))
IS_ANY_TOKEN = matcher.vocab.add_flag(lambda x: True)
matcher.add_pattern(
"FarAway",
[
{LOWER: "bob"},
{'OP': '*', LOWER: 'and'},
{LOWER: 'frank'}
])
doc = Doc(matcher.vocab, words=['bob', 'and', 'and', 'frank'])
match = matcher(doc)
assert len(match) == 1
ent_id, label, start, end = match[0]
assert start == 0
assert end == 4
@pytest.mark.xfail
def test_issue850():
'''The problem here is that the variable-length pattern matches the
succeeding token. We then don't handle the ambiguity correctly.'''
matcher = Matcher(Vocab(
lex_attr_getters={LOWER: lambda string: string.lower()}))
IS_ANY_TOKEN = matcher.vocab.add_flag(lambda x: True)
matcher.add_pattern(
"FarAway",
[
{LOWER: "bob"},
{'OP': '*', IS_ANY_TOKEN: True},
{LOWER: 'frank'}
])
doc = Doc(matcher.vocab, words=['bob', 'and', 'and', 'frank'])
match = matcher(doc)
assert len(match) == 1
ent_id, label, start, end = match[0]
assert start == 0
assert end == 4
| Update regression test for variable-length pattern problem in the matcher. | Update regression test for variable-length pattern problem in the matcher.
| Python | mit | aikramer2/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,spacy-io/spaCy,explosion/spaCy,oroszgy/spaCy.hu,oroszgy/spaCy.hu,explosion/spaCy,recognai/spaCy,raphael0202/spaCy,recognai/spaCy,honnibal/spaCy,raphael0202/spaCy,recognai/spaCy,honnibal/spaCy,aikramer2/spaCy,raphael0202/spaCy,explosion/spaCy,honnibal/spaCy,Gregory-Howard/spaCy,explosion/spaCy,recognai/spaCy,oroszgy/spaCy.hu,spacy-io/spaCy,spacy-io/spaCy,recognai/spaCy,recognai/spaCy,aikramer2/spaCy,spacy-io/spaCy,Gregory-Howard/spaCy,raphael0202/spaCy,explosion/spaCy,oroszgy/spaCy.hu,spacy-io/spaCy,raphael0202/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,explosion/spaCy,honnibal/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,Gregory-Howard/spaCy,oroszgy/spaCy.hu,aikramer2/spaCy,Gregory-Howard/spaCy,spacy-io/spaCy | '''
Test Matcher matches with '*' operator and Boolean flag
'''
from __future__ import unicode_literals
+ from __future__ import print_function
import pytest
from ...matcher import Matcher
from ...vocab import Vocab
from ...attrs import LOWER
from ...tokens import Doc
+ def test_basic_case():
+ matcher = Matcher(Vocab(
+ lex_attr_getters={LOWER: lambda string: string.lower()}))
+ IS_ANY_TOKEN = matcher.vocab.add_flag(lambda x: True)
+ matcher.add_pattern(
+ "FarAway",
+ [
+ {LOWER: "bob"},
+ {'OP': '*', LOWER: 'and'},
+ {LOWER: 'frank'}
+ ])
+ doc = Doc(matcher.vocab, words=['bob', 'and', 'and', 'frank'])
+ match = matcher(doc)
+ assert len(match) == 1
+ ent_id, label, start, end = match[0]
+ assert start == 0
+ assert end == 4
+
@pytest.mark.xfail
def test_issue850():
+ '''The problem here is that the variable-length pattern matches the
+ succeeding token. We then don't handle the ambiguity correctly.'''
- matcher = Matcher(Vocab())
+ matcher = Matcher(Vocab(
+ lex_attr_getters={LOWER: lambda string: string.lower()}))
IS_ANY_TOKEN = matcher.vocab.add_flag(lambda x: True)
matcher.add_pattern(
"FarAway",
[
{LOWER: "bob"},
{'OP': '*', IS_ANY_TOKEN: True},
{LOWER: 'frank'}
])
- doc = Doc(matcher.vocab, words=['bob', 'and', 'and', 'cat', 'frank'])
+ doc = Doc(matcher.vocab, words=['bob', 'and', 'and', 'frank'])
match = matcher(doc)
assert len(match) == 1
- start, end, label, ent_id = match
+ ent_id, label, start, end = match[0]
assert start == 0
assert end == 4
| Update regression test for variable-length pattern problem in the matcher. | ## Code Before:
'''
Test Matcher matches with '*' operator and Boolean flag
'''
from __future__ import unicode_literals
import pytest
from ...matcher import Matcher
from ...vocab import Vocab
from ...attrs import LOWER
from ...tokens import Doc
@pytest.mark.xfail
def test_issue850():
matcher = Matcher(Vocab())
IS_ANY_TOKEN = matcher.vocab.add_flag(lambda x: True)
matcher.add_pattern(
"FarAway",
[
{LOWER: "bob"},
{'OP': '*', IS_ANY_TOKEN: True},
{LOWER: 'frank'}
])
doc = Doc(matcher.vocab, words=['bob', 'and', 'and', 'cat', 'frank'])
match = matcher(doc)
assert len(match) == 1
start, end, label, ent_id = match
assert start == 0
assert end == 4
## Instruction:
Update regression test for variable-length pattern problem in the matcher.
## Code After:
'''
Test Matcher matches with '*' operator and Boolean flag
'''
from __future__ import unicode_literals
from __future__ import print_function
import pytest
from ...matcher import Matcher
from ...vocab import Vocab
from ...attrs import LOWER
from ...tokens import Doc
def test_basic_case():
matcher = Matcher(Vocab(
lex_attr_getters={LOWER: lambda string: string.lower()}))
IS_ANY_TOKEN = matcher.vocab.add_flag(lambda x: True)
matcher.add_pattern(
"FarAway",
[
{LOWER: "bob"},
{'OP': '*', LOWER: 'and'},
{LOWER: 'frank'}
])
doc = Doc(matcher.vocab, words=['bob', 'and', 'and', 'frank'])
match = matcher(doc)
assert len(match) == 1
ent_id, label, start, end = match[0]
assert start == 0
assert end == 4
@pytest.mark.xfail
def test_issue850():
'''The problem here is that the variable-length pattern matches the
succeeding token. We then don't handle the ambiguity correctly.'''
matcher = Matcher(Vocab(
lex_attr_getters={LOWER: lambda string: string.lower()}))
IS_ANY_TOKEN = matcher.vocab.add_flag(lambda x: True)
matcher.add_pattern(
"FarAway",
[
{LOWER: "bob"},
{'OP': '*', IS_ANY_TOKEN: True},
{LOWER: 'frank'}
])
doc = Doc(matcher.vocab, words=['bob', 'and', 'and', 'frank'])
match = matcher(doc)
assert len(match) == 1
ent_id, label, start, end = match[0]
assert start == 0
assert end == 4
|
fa8783f3307582dafcf636f5c94a7e4cff05724b | bin/tree_print_fasta_names.py | bin/tree_print_fasta_names.py |
import os
import shutil
import datetime
import sys
import argparse
from ete3 import Tree
import logging
DEFAULT_FORMAT = 1
class TreeIndex:
def __init__(self,tree_newick_fn,format=DEFAULT_FORMAT):
self.tree_newick_fn=tree_newick_fn
self.tree=read_newick(tree_newick_fn,format=format)
def process_node(self,node):
if node.is_leaf():
if hasattr(node,"fastapath"):
fastas_fn=node.fastapath.split("@")
for fasta_fn in fastas_fn:
print(fasta_fn)
else:
children=node.get_children()
for child in children:
self.process_node(child)
if __name__ == "__main__":
assert(len(sys.argv)==2)
newick_fn=sys.argv[1]
ti=TreeIndex(
tree_newick_fn=newick_fn,
)
ti.process_node(ti.tree.get_tree_root())
|
import os
import shutil
import datetime
import sys
from ete3 import Tree
DEFAULT_FORMAT = 1
class TreeIndex:
def __init__(self,tree_newick_fn,format=DEFAULT_FORMAT):
self.tree_newick_fn=tree_newick_fn
self.tree=Tree(tree_newick_fn,format=format)
def process_node(self,node):
if node.is_leaf():
if hasattr(node,"fastapath"):
fastas_fn=node.fastapath.split("@")
for fasta_fn in fastas_fn:
print(fasta_fn)
else:
children=node.get_children()
for child in children:
self.process_node(child)
if __name__ == "__main__":
assert(len(sys.argv)==2)
newick_fn=sys.argv[1]
ti=TreeIndex(
tree_newick_fn=newick_fn,
)
ti.process_node(ti.tree.get_tree_root())
| Fix error in loading trees | Fix error in loading trees
Former-commit-id: 6fda03a47c5fa2d65c143ebdd81e158ba5e1ccda | Python | mit | karel-brinda/prophyle,karel-brinda/prophyle,karel-brinda/prophyle,karel-brinda/prophyle |
import os
import shutil
import datetime
import sys
- import argparse
from ete3 import Tree
-
- import logging
DEFAULT_FORMAT = 1
class TreeIndex:
def __init__(self,tree_newick_fn,format=DEFAULT_FORMAT):
self.tree_newick_fn=tree_newick_fn
- self.tree=read_newick(tree_newick_fn,format=format)
+ self.tree=Tree(tree_newick_fn,format=format)
def process_node(self,node):
if node.is_leaf():
if hasattr(node,"fastapath"):
fastas_fn=node.fastapath.split("@")
for fasta_fn in fastas_fn:
print(fasta_fn)
else:
children=node.get_children()
for child in children:
self.process_node(child)
if __name__ == "__main__":
assert(len(sys.argv)==2)
newick_fn=sys.argv[1]
ti=TreeIndex(
tree_newick_fn=newick_fn,
)
ti.process_node(ti.tree.get_tree_root())
| Fix error in loading trees | ## Code Before:
import os
import shutil
import datetime
import sys
import argparse
from ete3 import Tree
import logging
DEFAULT_FORMAT = 1
class TreeIndex:
def __init__(self,tree_newick_fn,format=DEFAULT_FORMAT):
self.tree_newick_fn=tree_newick_fn
self.tree=read_newick(tree_newick_fn,format=format)
def process_node(self,node):
if node.is_leaf():
if hasattr(node,"fastapath"):
fastas_fn=node.fastapath.split("@")
for fasta_fn in fastas_fn:
print(fasta_fn)
else:
children=node.get_children()
for child in children:
self.process_node(child)
if __name__ == "__main__":
assert(len(sys.argv)==2)
newick_fn=sys.argv[1]
ti=TreeIndex(
tree_newick_fn=newick_fn,
)
ti.process_node(ti.tree.get_tree_root())
## Instruction:
Fix error in loading trees
## Code After:
import os
import shutil
import datetime
import sys
from ete3 import Tree
DEFAULT_FORMAT = 1
class TreeIndex:
def __init__(self,tree_newick_fn,format=DEFAULT_FORMAT):
self.tree_newick_fn=tree_newick_fn
self.tree=Tree(tree_newick_fn,format=format)
def process_node(self,node):
if node.is_leaf():
if hasattr(node,"fastapath"):
fastas_fn=node.fastapath.split("@")
for fasta_fn in fastas_fn:
print(fasta_fn)
else:
children=node.get_children()
for child in children:
self.process_node(child)
if __name__ == "__main__":
assert(len(sys.argv)==2)
newick_fn=sys.argv[1]
ti=TreeIndex(
tree_newick_fn=newick_fn,
)
ti.process_node(ti.tree.get_tree_root())
|
a20c6d072d70c535ed1f116fc04016c834ea9c14 | doc/en/_getdoctarget.py | doc/en/_getdoctarget.py |
import py
def get_version_string():
fn = py.path.local(__file__).join("..", "..", "..",
"_pytest", "__init__.py")
for line in fn.readlines():
if "version" in line:
return eval(line.split("=")[-1])
def get_minor_version_string():
return ".".join(get_version_string().split(".")[:2])
if __name__ == "__main__":
print (get_minor_version_string())
|
import py
def get_version_string():
fn = py.path.local(__file__).join("..", "..", "..",
"_pytest", "__init__.py")
for line in fn.readlines():
if "version" in line and not line.strip().startswith('#'):
return eval(line.split("=")[-1])
def get_minor_version_string():
return ".".join(get_version_string().split(".")[:2])
if __name__ == "__main__":
print (get_minor_version_string())
| Fix getdoctarget to ignore comment lines | Fix getdoctarget to ignore comment lines
| Python | mit | etataurov/pytest,gabrielcnr/pytest,mbirtwell/pytest,vodik/pytest,The-Compiler/pytest,omarkohl/pytest,Bjwebb/pytest,davidszotten/pytest,gabrielcnr/pytest,mdboom/pytest,ionelmc/pytest,malinoff/pytest,hpk42/pytest,tareqalayan/pytest,userzimmermann/pytest,rouge8/pytest,tgoodlet/pytest,abusalimov/pytest,bukzor/pytest,icemac/pytest,pfctdayelise/pytest,JonathonSonesen/pytest,ionelmc/pytest,alfredodeza/pytest,chiller/pytest,skylarjhdownes/pytest,RonnyPfannschmidt/pytest,Haibo-Wang-ORG/pytest,lukas-bednar/pytest,mhils/pytest,mhils/pytest,chiller/pytest,oleg-alexandrov/pytest,oleg-alexandrov/pytest,MengJueM/pytest,chillbear/pytest,rmfitzpatrick/pytest,tomviner/pytest,Bachmann1234/pytest,pytest-dev/pytest,doordash/pytest,eli-b/pytest,codewarrior0/pytest,flub/pytest,bukzor/pytest,ropez/pytest,Haibo-Wang-ORG/pytest,abusalimov/pytest,mdboom/pytest,MengJueM/pytest,icemac/pytest,The-Compiler/pytest,vmalloc/dessert,codewarrior0/pytest,jb098/pytest,chillbear/pytest,jb098/pytest,omarkohl/pytest,doordash/pytest,mbirtwell/pytest,nicoddemus/pytest,tomviner/pytest,nicoddemus/pytest,ropez/pytest,jaraco/pytest,rouge8/pytest,markshao/pytest,txomon/pytest,lukas-bednar/pytest,Bachmann1234/pytest,userzimmermann/pytest,MichaelAquilina/pytest,vodik/pytest,hpk42/pytest,ddboline/pytest,hackebrot/pytest,JonathonSonesen/pytest,Akasurde/pytest,Bjwebb/pytest |
import py
def get_version_string():
fn = py.path.local(__file__).join("..", "..", "..",
"_pytest", "__init__.py")
for line in fn.readlines():
- if "version" in line:
+ if "version" in line and not line.strip().startswith('#'):
return eval(line.split("=")[-1])
def get_minor_version_string():
return ".".join(get_version_string().split(".")[:2])
if __name__ == "__main__":
print (get_minor_version_string())
| Fix getdoctarget to ignore comment lines | ## Code Before:
import py
def get_version_string():
fn = py.path.local(__file__).join("..", "..", "..",
"_pytest", "__init__.py")
for line in fn.readlines():
if "version" in line:
return eval(line.split("=")[-1])
def get_minor_version_string():
return ".".join(get_version_string().split(".")[:2])
if __name__ == "__main__":
print (get_minor_version_string())
## Instruction:
Fix getdoctarget to ignore comment lines
## Code After:
import py
def get_version_string():
fn = py.path.local(__file__).join("..", "..", "..",
"_pytest", "__init__.py")
for line in fn.readlines():
if "version" in line and not line.strip().startswith('#'):
return eval(line.split("=")[-1])
def get_minor_version_string():
return ".".join(get_version_string().split(".")[:2])
if __name__ == "__main__":
print (get_minor_version_string())
|
efd6fad89131c4d3070c68013ace77f11647bd68 | opal/core/search/__init__.py | opal/core/search/__init__.py | from opal.core.search import urls
from opal.core import plugins
from opal.core import celery # NOQA
class SearchPlugin(plugins.OpalPlugin):
"""
The plugin entrypoint for OPAL's core search functionality
"""
urls = urls.urlpatterns
javascripts = {
'opal.services': [
'js/search/services/filter.js',
'js/search/services/filters_loader.js',
'js/search/services/filter_resource.js',
"js/search/services/paginator.js",
],
'opal.controllers': [
'js/search/controllers/search.js',
'js/search/controllers/extract.js',
"js/search/controllers/save_filter.js",
]
}
plugins.register(SearchPlugin)
| from opal.core import celery # NOQA
from opal.core.search import plugin
| Move Opal.core.search plugin into a plugins.py ahead of full plugin 2.0 refactor | Move Opal.core.search plugin into a plugins.py ahead of full plugin 2.0 refactor
| Python | agpl-3.0 | khchine5/opal,khchine5/opal,khchine5/opal | + from opal.core import celery # NOQA
- from opal.core.search import urls
+ from opal.core.search import plugin
- from opal.core import plugins
-
- from opal.core import celery # NOQA
-
-
- class SearchPlugin(plugins.OpalPlugin):
- """
- The plugin entrypoint for OPAL's core search functionality
- """
- urls = urls.urlpatterns
- javascripts = {
- 'opal.services': [
- 'js/search/services/filter.js',
- 'js/search/services/filters_loader.js',
- 'js/search/services/filter_resource.js',
- "js/search/services/paginator.js",
- ],
- 'opal.controllers': [
- 'js/search/controllers/search.js',
- 'js/search/controllers/extract.js',
- "js/search/controllers/save_filter.js",
- ]
- }
-
- plugins.register(SearchPlugin)
- | Move Opal.core.search plugin into a plugins.py ahead of full plugin 2.0 refactor | ## Code Before:
from opal.core.search import urls
from opal.core import plugins
from opal.core import celery # NOQA
class SearchPlugin(plugins.OpalPlugin):
"""
The plugin entrypoint for OPAL's core search functionality
"""
urls = urls.urlpatterns
javascripts = {
'opal.services': [
'js/search/services/filter.js',
'js/search/services/filters_loader.js',
'js/search/services/filter_resource.js',
"js/search/services/paginator.js",
],
'opal.controllers': [
'js/search/controllers/search.js',
'js/search/controllers/extract.js',
"js/search/controllers/save_filter.js",
]
}
plugins.register(SearchPlugin)
## Instruction:
Move Opal.core.search plugin into a plugins.py ahead of full plugin 2.0 refactor
## Code After:
from opal.core import celery # NOQA
from opal.core.search import plugin
|
56a8b900570200e63ee460dd7e2962cba2450b16 | preparation/tools/build_assets.py | preparation/tools/build_assets.py | from copy import copy
import argparse
from preparation.resources.Resource import names_registered, resource_by_name
from hb_res.storage import get_storage, ExplanationStorage
def generate_asset(resource, out_storage: ExplanationStorage):
out_storage.clear()
for explanation in resource:
r = copy(explanation)
for functor in resource.modifiers:
if r is None:
break
r = functor(r)
if r is not None:
out_storage.add_entry(r)
def rebuild_trunk(trunk: str):
resource = resource_by_name(trunk + 'Resource')()
with get_storage(trunk) as out_storage:
print("Starting {} generation".format(trunk))
generate_asset(resource, out_storage)
print("Finished {} generation".format(trunk))
def make_argparser():
parser = argparse.ArgumentParser(description='Rebuild some asset')
names = [name.replace('Resource', '') for name in names_registered()]
parser.add_argument('resources',
metavar='RESOURCE',
nargs='+',
choices=names + ['all'],
help='One of registered resources ({}) or just \'all\'.'.format(', '.join(names)))
return parser
def main(args=None):
if not isinstance(args, argparse.Namespace):
parser = make_argparser()
args = parser.parse_args(args)
assert all not in args.resources or len(args.resources) == 1
for name in args.resources:
rebuild_trunk(name)
if __name__ == '__main__':
main()
| from copy import copy
import argparse
from preparation.resources.Resource import names_registered, resource_by_name
from hb_res.storage import get_storage, ExplanationStorage
def generate_asset(resource, out_storage: ExplanationStorage):
out_storage.clear()
for explanation in resource:
r = copy(explanation)
for functor in resource.modifiers:
if r is None:
break
r = functor(r)
if r is not None:
out_storage.add_entry(r)
def rebuild_trunk(trunk: str):
resource = resource_by_name(trunk + 'Resource')()
with get_storage(trunk) as out_storage:
print("Starting {} generation".format(trunk))
generate_asset(resource, out_storage)
print("Finished {} generation".format(trunk))
def make_argparser():
parser = argparse.ArgumentParser(description='Rebuild some asset')
names = [name.replace('Resource', '') for name in names_registered()]
parser.add_argument('resources',
metavar='RESOURCE',
nargs='+',
choices=names + ['all'],
help='One of registered resources ({}) or just \'all\'.'.format(', '.join(names)))
return parser
def main(args=None):
if not isinstance(args, argparse.Namespace):
parser = make_argparser()
args = parser.parse_args(args)
assert 'all' not in args.resources or len(args.resources) == 1
if 'all' in args.resources:
args.resources = [name.replace('Resource', '') for name in names_registered()]
for name in args.resources:
rebuild_trunk(name)
if __name__ == '__main__':
main()
| Fix bug with 'all' argument | Fix bug with 'all' argument
| Python | mit | hatbot-team/hatbot_resources | from copy import copy
import argparse
from preparation.resources.Resource import names_registered, resource_by_name
from hb_res.storage import get_storage, ExplanationStorage
def generate_asset(resource, out_storage: ExplanationStorage):
out_storage.clear()
for explanation in resource:
r = copy(explanation)
for functor in resource.modifiers:
if r is None:
break
r = functor(r)
if r is not None:
out_storage.add_entry(r)
def rebuild_trunk(trunk: str):
resource = resource_by_name(trunk + 'Resource')()
with get_storage(trunk) as out_storage:
print("Starting {} generation".format(trunk))
generate_asset(resource, out_storage)
print("Finished {} generation".format(trunk))
def make_argparser():
parser = argparse.ArgumentParser(description='Rebuild some asset')
names = [name.replace('Resource', '') for name in names_registered()]
parser.add_argument('resources',
metavar='RESOURCE',
nargs='+',
choices=names + ['all'],
help='One of registered resources ({}) or just \'all\'.'.format(', '.join(names)))
return parser
def main(args=None):
if not isinstance(args, argparse.Namespace):
parser = make_argparser()
args = parser.parse_args(args)
- assert all not in args.resources or len(args.resources) == 1
+ assert 'all' not in args.resources or len(args.resources) == 1
+ if 'all' in args.resources:
+ args.resources = [name.replace('Resource', '') for name in names_registered()]
for name in args.resources:
rebuild_trunk(name)
if __name__ == '__main__':
main()
| Fix bug with 'all' argument | ## Code Before:
from copy import copy
import argparse
from preparation.resources.Resource import names_registered, resource_by_name
from hb_res.storage import get_storage, ExplanationStorage
def generate_asset(resource, out_storage: ExplanationStorage):
out_storage.clear()
for explanation in resource:
r = copy(explanation)
for functor in resource.modifiers:
if r is None:
break
r = functor(r)
if r is not None:
out_storage.add_entry(r)
def rebuild_trunk(trunk: str):
resource = resource_by_name(trunk + 'Resource')()
with get_storage(trunk) as out_storage:
print("Starting {} generation".format(trunk))
generate_asset(resource, out_storage)
print("Finished {} generation".format(trunk))
def make_argparser():
parser = argparse.ArgumentParser(description='Rebuild some asset')
names = [name.replace('Resource', '') for name in names_registered()]
parser.add_argument('resources',
metavar='RESOURCE',
nargs='+',
choices=names + ['all'],
help='One of registered resources ({}) or just \'all\'.'.format(', '.join(names)))
return parser
def main(args=None):
if not isinstance(args, argparse.Namespace):
parser = make_argparser()
args = parser.parse_args(args)
assert all not in args.resources or len(args.resources) == 1
for name in args.resources:
rebuild_trunk(name)
if __name__ == '__main__':
main()
## Instruction:
Fix bug with 'all' argument
## Code After:
from copy import copy
import argparse
from preparation.resources.Resource import names_registered, resource_by_name
from hb_res.storage import get_storage, ExplanationStorage
def generate_asset(resource, out_storage: ExplanationStorage):
out_storage.clear()
for explanation in resource:
r = copy(explanation)
for functor in resource.modifiers:
if r is None:
break
r = functor(r)
if r is not None:
out_storage.add_entry(r)
def rebuild_trunk(trunk: str):
resource = resource_by_name(trunk + 'Resource')()
with get_storage(trunk) as out_storage:
print("Starting {} generation".format(trunk))
generate_asset(resource, out_storage)
print("Finished {} generation".format(trunk))
def make_argparser():
parser = argparse.ArgumentParser(description='Rebuild some asset')
names = [name.replace('Resource', '') for name in names_registered()]
parser.add_argument('resources',
metavar='RESOURCE',
nargs='+',
choices=names + ['all'],
help='One of registered resources ({}) or just \'all\'.'.format(', '.join(names)))
return parser
def main(args=None):
if not isinstance(args, argparse.Namespace):
parser = make_argparser()
args = parser.parse_args(args)
assert 'all' not in args.resources or len(args.resources) == 1
if 'all' in args.resources:
args.resources = [name.replace('Resource', '') for name in names_registered()]
for name in args.resources:
rebuild_trunk(name)
if __name__ == '__main__':
main()
|
6167215e4ed49e8a4300f327d5b4ed4540d1a420 | numba/tests/npyufunc/test_parallel_env_variable.py | numba/tests/npyufunc/test_parallel_env_variable.py | from numba.np.ufunc.parallel import get_thread_count
from os import environ as env
from numba.core import config
import unittest
class TestParallelEnvVariable(unittest.TestCase):
"""
Tests environment variables related to the underlying "parallel"
functions for npyufuncs.
"""
_numba_parallel_test_ = False
def test_num_threads_variable(self):
"""
Tests the NUMBA_NUM_THREADS env variable behaves as expected.
"""
key = 'NUMBA_NUM_THREADS'
current = str(getattr(env, key, config.NUMBA_DEFAULT_NUM_THREADS))
threads = "3154"
env[key] = threads
try:
config.reload_config()
except RuntimeError as e:
# This test should fail if threads have already been launched
self.assertIn("Cannot set NUMBA_NUM_THREADS", e.args[0])
else:
try:
self.assertEqual(threads, str(get_thread_count()))
self.assertEqual(threads, str(config.NUMBA_NUM_THREADS))
finally:
# reset the env variable/set to default
env[key] = current
config.reload_config()
if __name__ == '__main__':
unittest.main()
| from numba.np.ufunc.parallel import get_thread_count
from os import environ as env
from numba.core import config
import unittest
class TestParallelEnvVariable(unittest.TestCase):
"""
Tests environment variables related to the underlying "parallel"
functions for npyufuncs.
"""
_numba_parallel_test_ = False
def test_num_threads_variable(self):
"""
Tests the NUMBA_NUM_THREADS env variable behaves as expected.
"""
key = 'NUMBA_NUM_THREADS'
current = str(getattr(env, key, config.NUMBA_DEFAULT_NUM_THREADS))
threads = "3154"
env[key] = threads
try:
config.reload_config()
except RuntimeError as e:
# This test should fail if threads have already been launched
self.assertIn("Cannot set NUMBA_NUM_THREADS", e.args[0])
else:
self.assertEqual(threads, str(get_thread_count()))
self.assertEqual(threads, str(config.NUMBA_NUM_THREADS))
finally:
# reset the env variable/set to default. Should not fail even if
# threads are launched because the value is the same.
env[key] = current
config.reload_config()
if __name__ == '__main__':
unittest.main()
| Fix the parallel env variable test to reset the env correctly | Fix the parallel env variable test to reset the env correctly
| Python | bsd-2-clause | seibert/numba,seibert/numba,stonebig/numba,cpcloud/numba,sklam/numba,gmarkall/numba,stuartarchibald/numba,numba/numba,stonebig/numba,seibert/numba,stuartarchibald/numba,sklam/numba,gmarkall/numba,sklam/numba,sklam/numba,stonebig/numba,IntelLabs/numba,IntelLabs/numba,stuartarchibald/numba,cpcloud/numba,numba/numba,stuartarchibald/numba,IntelLabs/numba,stuartarchibald/numba,cpcloud/numba,gmarkall/numba,numba/numba,seibert/numba,IntelLabs/numba,gmarkall/numba,numba/numba,numba/numba,sklam/numba,gmarkall/numba,stonebig/numba,cpcloud/numba,stonebig/numba,IntelLabs/numba,seibert/numba,cpcloud/numba | from numba.np.ufunc.parallel import get_thread_count
from os import environ as env
from numba.core import config
import unittest
class TestParallelEnvVariable(unittest.TestCase):
"""
Tests environment variables related to the underlying "parallel"
functions for npyufuncs.
"""
_numba_parallel_test_ = False
def test_num_threads_variable(self):
"""
Tests the NUMBA_NUM_THREADS env variable behaves as expected.
"""
key = 'NUMBA_NUM_THREADS'
current = str(getattr(env, key, config.NUMBA_DEFAULT_NUM_THREADS))
threads = "3154"
env[key] = threads
try:
config.reload_config()
except RuntimeError as e:
# This test should fail if threads have already been launched
self.assertIn("Cannot set NUMBA_NUM_THREADS", e.args[0])
else:
- try:
- self.assertEqual(threads, str(get_thread_count()))
+ self.assertEqual(threads, str(get_thread_count()))
- self.assertEqual(threads, str(config.NUMBA_NUM_THREADS))
+ self.assertEqual(threads, str(config.NUMBA_NUM_THREADS))
- finally:
+ finally:
- # reset the env variable/set to default
+ # reset the env variable/set to default. Should not fail even if
+ # threads are launched because the value is the same.
- env[key] = current
+ env[key] = current
- config.reload_config()
+ config.reload_config()
if __name__ == '__main__':
unittest.main()
| Fix the parallel env variable test to reset the env correctly | ## Code Before:
from numba.np.ufunc.parallel import get_thread_count
from os import environ as env
from numba.core import config
import unittest
class TestParallelEnvVariable(unittest.TestCase):
"""
Tests environment variables related to the underlying "parallel"
functions for npyufuncs.
"""
_numba_parallel_test_ = False
def test_num_threads_variable(self):
"""
Tests the NUMBA_NUM_THREADS env variable behaves as expected.
"""
key = 'NUMBA_NUM_THREADS'
current = str(getattr(env, key, config.NUMBA_DEFAULT_NUM_THREADS))
threads = "3154"
env[key] = threads
try:
config.reload_config()
except RuntimeError as e:
# This test should fail if threads have already been launched
self.assertIn("Cannot set NUMBA_NUM_THREADS", e.args[0])
else:
try:
self.assertEqual(threads, str(get_thread_count()))
self.assertEqual(threads, str(config.NUMBA_NUM_THREADS))
finally:
# reset the env variable/set to default
env[key] = current
config.reload_config()
if __name__ == '__main__':
unittest.main()
## Instruction:
Fix the parallel env variable test to reset the env correctly
## Code After:
from numba.np.ufunc.parallel import get_thread_count
from os import environ as env
from numba.core import config
import unittest
class TestParallelEnvVariable(unittest.TestCase):
"""
Tests environment variables related to the underlying "parallel"
functions for npyufuncs.
"""
_numba_parallel_test_ = False
def test_num_threads_variable(self):
"""
Tests the NUMBA_NUM_THREADS env variable behaves as expected.
"""
key = 'NUMBA_NUM_THREADS'
current = str(getattr(env, key, config.NUMBA_DEFAULT_NUM_THREADS))
threads = "3154"
env[key] = threads
try:
config.reload_config()
except RuntimeError as e:
# This test should fail if threads have already been launched
self.assertIn("Cannot set NUMBA_NUM_THREADS", e.args[0])
else:
self.assertEqual(threads, str(get_thread_count()))
self.assertEqual(threads, str(config.NUMBA_NUM_THREADS))
finally:
# reset the env variable/set to default. Should not fail even if
# threads are launched because the value is the same.
env[key] = current
config.reload_config()
if __name__ == '__main__':
unittest.main()
|
504205d02ef2f5b66da225390fdb34b8b736ce57 | ideascube/migrations/0009_add_a_system_user.py | ideascube/migrations/0009_add_a_system_user.py | from __future__ import unicode_literals
from django.contrib.auth import get_user_model
from django.db import migrations
def add_user(*args):
User = get_user_model()
User(serial='__system__', full_name='System', password='!!').save()
class Migration(migrations.Migration):
dependencies = [
('ideascube', '0008_user_sdb_level'),
('search', '0001_initial'),
]
operations = [
migrations.RunPython(add_user, None),
]
| from __future__ import unicode_literals
from django.db import migrations
def add_user(apps, *args):
User = apps.get_model('ideascube', 'User')
User(serial='__system__', full_name='System', password='!!').save()
class Migration(migrations.Migration):
dependencies = [
('ideascube', '0008_user_sdb_level'),
('search', '0001_initial'),
]
operations = [
migrations.RunPython(add_user, None),
]
| Load user from migration registry when creating system user | Load user from migration registry when creating system user
Always load models from the registry in migration files.
I hate the idea of touching a migration already released, but
this one prevents us from adding new properties to User.
If we load the User directly (not from registry) when creating
the user model, we'll try to create a user with column that does
not exist at the time of this migration.
| Python | agpl-3.0 | ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube | from __future__ import unicode_literals
- from django.contrib.auth import get_user_model
from django.db import migrations
- def add_user(*args):
+ def add_user(apps, *args):
- User = get_user_model()
+ User = apps.get_model('ideascube', 'User')
User(serial='__system__', full_name='System', password='!!').save()
class Migration(migrations.Migration):
dependencies = [
('ideascube', '0008_user_sdb_level'),
('search', '0001_initial'),
]
operations = [
migrations.RunPython(add_user, None),
]
| Load user from migration registry when creating system user | ## Code Before:
from __future__ import unicode_literals
from django.contrib.auth import get_user_model
from django.db import migrations
def add_user(*args):
User = get_user_model()
User(serial='__system__', full_name='System', password='!!').save()
class Migration(migrations.Migration):
dependencies = [
('ideascube', '0008_user_sdb_level'),
('search', '0001_initial'),
]
operations = [
migrations.RunPython(add_user, None),
]
## Instruction:
Load user from migration registry when creating system user
## Code After:
from __future__ import unicode_literals
from django.db import migrations
def add_user(apps, *args):
User = apps.get_model('ideascube', 'User')
User(serial='__system__', full_name='System', password='!!').save()
class Migration(migrations.Migration):
dependencies = [
('ideascube', '0008_user_sdb_level'),
('search', '0001_initial'),
]
operations = [
migrations.RunPython(add_user, None),
]
|
d8d2e4b763fbd7cedc42046f6f45395bf15caa79 | samples/plugins/scenario/scenario_plugin.py | samples/plugins/scenario/scenario_plugin.py |
from rally.task.scenarios import base
class ScenarioPlugin(base.Scenario):
"""Sample plugin which lists flavors."""
@base.atomic_action_timer("list_flavors")
def _list_flavors(self):
"""Sample of usage clients - list flavors
You can use self.context, self.admin_clients and self.clients which are
initialized on scenario instance creation.
"""
self.clients("nova").flavors.list()
@base.atomic_action_timer("list_flavors_as_admin")
def _list_flavors_as_admin(self):
"""The same with admin clients."""
self.admin_clients("nova").flavors.list()
@base.scenario()
def list_flavors(self):
"""List flavors."""
self._list_flavors()
self._list_flavors_as_admin()
|
from rally.plugins.openstack import scenario
from rally.task import atomic
class ScenarioPlugin(scenario.OpenStackScenario):
"""Sample plugin which lists flavors."""
@atomic.action_timer("list_flavors")
def _list_flavors(self):
"""Sample of usage clients - list flavors
You can use self.context, self.admin_clients and self.clients which are
initialized on scenario instance creation.
"""
self.clients("nova").flavors.list()
@atomic.action_timer("list_flavors_as_admin")
def _list_flavors_as_admin(self):
"""The same with admin clients."""
self.admin_clients("nova").flavors.list()
@scenario.configure()
def list_flavors(self):
"""List flavors."""
self._list_flavors()
self._list_flavors_as_admin()
| Fix the scenario plugin sample | Fix the scenario plugin sample
We forgot to fix scenario plugin sample when we were doing
rally.task.scenario refactoring
Change-Id: Iadbb960cf168bd3b9cd6c1881a5f7a8dffd7036f
| Python | apache-2.0 | group-policy/rally,eayunstack/rally,openstack/rally,amit0701/rally,paboldin/rally,eonpatapon/rally,cernops/rally,afaheem88/rally,yeming233/rally,vganapath/rally,gluke77/rally,aforalee/RRally,yeming233/rally,openstack/rally,gluke77/rally,gluke77/rally,eayunstack/rally,openstack/rally,aforalee/RRally,cernops/rally,group-policy/rally,openstack/rally,vganapath/rally,amit0701/rally,vganapath/rally,redhat-openstack/rally,eonpatapon/rally,amit0701/rally,gluke77/rally,paboldin/rally,redhat-openstack/rally,paboldin/rally,group-policy/rally,afaheem88/rally,eayunstack/rally,aplanas/rally,vganapath/rally,aplanas/rally |
- from rally.task.scenarios import base
+ from rally.plugins.openstack import scenario
+ from rally.task import atomic
- class ScenarioPlugin(base.Scenario):
+ class ScenarioPlugin(scenario.OpenStackScenario):
"""Sample plugin which lists flavors."""
- @base.atomic_action_timer("list_flavors")
+ @atomic.action_timer("list_flavors")
def _list_flavors(self):
"""Sample of usage clients - list flavors
You can use self.context, self.admin_clients and self.clients which are
initialized on scenario instance creation.
"""
self.clients("nova").flavors.list()
- @base.atomic_action_timer("list_flavors_as_admin")
+ @atomic.action_timer("list_flavors_as_admin")
def _list_flavors_as_admin(self):
"""The same with admin clients."""
self.admin_clients("nova").flavors.list()
- @base.scenario()
+ @scenario.configure()
def list_flavors(self):
"""List flavors."""
self._list_flavors()
self._list_flavors_as_admin()
| Fix the scenario plugin sample | ## Code Before:
from rally.task.scenarios import base
class ScenarioPlugin(base.Scenario):
"""Sample plugin which lists flavors."""
@base.atomic_action_timer("list_flavors")
def _list_flavors(self):
"""Sample of usage clients - list flavors
You can use self.context, self.admin_clients and self.clients which are
initialized on scenario instance creation.
"""
self.clients("nova").flavors.list()
@base.atomic_action_timer("list_flavors_as_admin")
def _list_flavors_as_admin(self):
"""The same with admin clients."""
self.admin_clients("nova").flavors.list()
@base.scenario()
def list_flavors(self):
"""List flavors."""
self._list_flavors()
self._list_flavors_as_admin()
## Instruction:
Fix the scenario plugin sample
## Code After:
from rally.plugins.openstack import scenario
from rally.task import atomic
class ScenarioPlugin(scenario.OpenStackScenario):
"""Sample plugin which lists flavors."""
@atomic.action_timer("list_flavors")
def _list_flavors(self):
"""Sample of usage clients - list flavors
You can use self.context, self.admin_clients and self.clients which are
initialized on scenario instance creation.
"""
self.clients("nova").flavors.list()
@atomic.action_timer("list_flavors_as_admin")
def _list_flavors_as_admin(self):
"""The same with admin clients."""
self.admin_clients("nova").flavors.list()
@scenario.configure()
def list_flavors(self):
"""List flavors."""
self._list_flavors()
self._list_flavors_as_admin()
|
68a621005c5a520b7a97c4cad462d43fb7f3aaed | paws/views.py | paws/views.py |
from .request import Request
from .response import Response, response
import logging
log = logging.getLogger()
class View:
def __call__(self, event, context):
request = Request(event, context)
resp = self.prepare(request)
if resp:
return resp
kwargs = event.get('pathParameters') or {}
func = getattr(self, request.method.lower())
try:
resp = func(request, **kwargs)
except:
import traceback
log.error(self)
log.error(traceback.format_exc())
return response(body='Internal server Error', status=500)
if isinstance(resp, Response):
resp = resp.render()
return resp
def prepare(self, request):
pass
|
from .request import Request
from .response import Response, response
import logging
log = logging.getLogger()
class View:
def __call__(self, event, context):
kwargs = event.get('pathParameters') or {}
self.dispatch(request, **kwargs)
def dispatch(self, request, **kwargs):
func = getattr(self, request.method.lower())
try:
resp = func(request, **kwargs)
except:
import traceback
log.error(self)
log.error(traceback.format_exc())
return response(body='Internal server Error', status=500)
if isinstance(resp, Response):
resp = resp.render()
return resp
def prepare(self, request):
pass
| Break out dispatch, and drop prepare. Easier testing | Break out dispatch, and drop prepare. Easier testing
| Python | bsd-3-clause | funkybob/paws |
from .request import Request
from .response import Response, response
import logging
log = logging.getLogger()
class View:
def __call__(self, event, context):
- request = Request(event, context)
- resp = self.prepare(request)
- if resp:
- return resp
kwargs = event.get('pathParameters') or {}
+ self.dispatch(request, **kwargs)
+
+ def dispatch(self, request, **kwargs):
func = getattr(self, request.method.lower())
try:
resp = func(request, **kwargs)
except:
import traceback
log.error(self)
log.error(traceback.format_exc())
return response(body='Internal server Error', status=500)
if isinstance(resp, Response):
resp = resp.render()
return resp
def prepare(self, request):
pass
| Break out dispatch, and drop prepare. Easier testing | ## Code Before:
from .request import Request
from .response import Response, response
import logging
log = logging.getLogger()
class View:
def __call__(self, event, context):
request = Request(event, context)
resp = self.prepare(request)
if resp:
return resp
kwargs = event.get('pathParameters') or {}
func = getattr(self, request.method.lower())
try:
resp = func(request, **kwargs)
except:
import traceback
log.error(self)
log.error(traceback.format_exc())
return response(body='Internal server Error', status=500)
if isinstance(resp, Response):
resp = resp.render()
return resp
def prepare(self, request):
pass
## Instruction:
Break out dispatch, and drop prepare. Easier testing
## Code After:
from .request import Request
from .response import Response, response
import logging
log = logging.getLogger()
class View:
def __call__(self, event, context):
kwargs = event.get('pathParameters') or {}
self.dispatch(request, **kwargs)
def dispatch(self, request, **kwargs):
func = getattr(self, request.method.lower())
try:
resp = func(request, **kwargs)
except:
import traceback
log.error(self)
log.error(traceback.format_exc())
return response(body='Internal server Error', status=500)
if isinstance(resp, Response):
resp = resp.render()
return resp
def prepare(self, request):
pass
|
d4d448adff71b609d5efb269d1a9a2ea4aba3590 | radio/templatetags/radio_js_config.py | radio/templatetags/radio_js_config.py | import random
import json
from django import template
from django.conf import settings
register = template.Library()
# Build json value to pass as js config
@register.simple_tag()
def trunkplayer_js_config(user):
js_settings = getattr(settings, 'JS_SETTINGS', None)
js_json = {}
if js_settings:
for setting in js_settings:
set_val = getattr(settings, setting, '')
js_json[setting] = set_val
js_json['user_is_staff'] = user.is_staff
if user.is_authenticated():
js_json['user_is_authenticated'] = True
else:
js_json['user_is_authenticated'] = False
js_json['radio_change_unit'] = user.has_perm('radio.change_unit')
return json.dumps(js_json)
| import random
import json
from django import template
from django.conf import settings
from radio.models import SiteOption
register = template.Library()
# Build json value to pass as js config
@register.simple_tag()
def trunkplayer_js_config(user):
js_settings = getattr(settings, 'JS_SETTINGS', None)
js_json = {}
if js_settings:
for setting in js_settings:
set_val = getattr(settings, setting, '')
js_json[setting] = set_val
for opt in SiteOption.objects.filter(javascript_visible=True):
js_json[opt.name] = opt.value_boolean_or_string()
js_json['user_is_staff'] = user.is_staff
if user.is_authenticated():
js_json['user_is_authenticated'] = True
else:
js_json['user_is_authenticated'] = False
js_json['radio_change_unit'] = user.has_perm('radio.change_unit')
return json.dumps(js_json)
| Allow SiteOption to load into the JS | Allow SiteOption to load into the JS
| Python | mit | ScanOC/trunk-player,ScanOC/trunk-player,ScanOC/trunk-player,ScanOC/trunk-player | import random
import json
from django import template
from django.conf import settings
+
+ from radio.models import SiteOption
register = template.Library()
# Build json value to pass as js config
@register.simple_tag()
def trunkplayer_js_config(user):
js_settings = getattr(settings, 'JS_SETTINGS', None)
js_json = {}
if js_settings:
for setting in js_settings:
set_val = getattr(settings, setting, '')
js_json[setting] = set_val
+ for opt in SiteOption.objects.filter(javascript_visible=True):
+ js_json[opt.name] = opt.value_boolean_or_string()
js_json['user_is_staff'] = user.is_staff
if user.is_authenticated():
js_json['user_is_authenticated'] = True
else:
js_json['user_is_authenticated'] = False
js_json['radio_change_unit'] = user.has_perm('radio.change_unit')
return json.dumps(js_json)
| Allow SiteOption to load into the JS | ## Code Before:
import random
import json
from django import template
from django.conf import settings
register = template.Library()
# Build json value to pass as js config
@register.simple_tag()
def trunkplayer_js_config(user):
js_settings = getattr(settings, 'JS_SETTINGS', None)
js_json = {}
if js_settings:
for setting in js_settings:
set_val = getattr(settings, setting, '')
js_json[setting] = set_val
js_json['user_is_staff'] = user.is_staff
if user.is_authenticated():
js_json['user_is_authenticated'] = True
else:
js_json['user_is_authenticated'] = False
js_json['radio_change_unit'] = user.has_perm('radio.change_unit')
return json.dumps(js_json)
## Instruction:
Allow SiteOption to load into the JS
## Code After:
import random
import json
from django import template
from django.conf import settings
from radio.models import SiteOption
register = template.Library()
# Build json value to pass as js config
@register.simple_tag()
def trunkplayer_js_config(user):
js_settings = getattr(settings, 'JS_SETTINGS', None)
js_json = {}
if js_settings:
for setting in js_settings:
set_val = getattr(settings, setting, '')
js_json[setting] = set_val
for opt in SiteOption.objects.filter(javascript_visible=True):
js_json[opt.name] = opt.value_boolean_or_string()
js_json['user_is_staff'] = user.is_staff
if user.is_authenticated():
js_json['user_is_authenticated'] = True
else:
js_json['user_is_authenticated'] = False
js_json['radio_change_unit'] = user.has_perm('radio.change_unit')
return json.dumps(js_json)
|
0a13a9a8a779102dbcb2beead7d8aa9143f4c79b | tests/pytests/unit/client/ssh/test_shell.py | tests/pytests/unit/client/ssh/test_shell.py | import os
import subprocess
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
yield {"pub_key": str(pub_key), "priv_key": str(priv_key)}
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
class TestSSHShell:
def test_ssh_shell_key_gen(self, keys):
"""
Test ssh key_gen
"""
shell.gen_key(keys["priv_key"])
for fp in keys.keys():
assert os.path.exists(keys[fp])
# verify there is not a passphrase set on key
ret = subprocess.check_output(
["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30,
)
assert ret.decode().startswith("ssh-rsa")
| import subprocess
import types
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key)
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
def test_ssh_shell_key_gen(keys):
"""
Test ssh key_gen
"""
shell.gen_key(str(keys.priv_key))
assert keys.priv_key.exists()
assert keys.pub_key.exists()
# verify there is not a passphrase set on key
ret = subprocess.check_output(
["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30,
)
assert ret.decode().startswith("ssh-rsa")
| Use commit suggestion to use types | Use commit suggestion to use types
Co-authored-by: Pedro Algarvio <4410d99cefe57ec2c2cdbd3f1d5cf862bb4fb6f8@algarvio.me>
| Python | apache-2.0 | saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt | - import os
import subprocess
+ import types
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
- yield {"pub_key": str(pub_key), "priv_key": str(priv_key)}
+ return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key)
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
- class TestSSHShell:
- def test_ssh_shell_key_gen(self, keys):
+ def test_ssh_shell_key_gen(keys):
- """
+ """
- Test ssh key_gen
+ Test ssh key_gen
- """
+ """
- shell.gen_key(keys["priv_key"])
+ shell.gen_key(str(keys.priv_key))
- for fp in keys.keys():
- assert os.path.exists(keys[fp])
+ assert keys.priv_key.exists()
+ assert keys.pub_key.exists()
+ # verify there is not a passphrase set on key
+ ret = subprocess.check_output(
+ ["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30,
+ )
+ assert ret.decode().startswith("ssh-rsa")
- # verify there is not a passphrase set on key
- ret = subprocess.check_output(
- ["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30,
- )
- assert ret.decode().startswith("ssh-rsa")
- | Use commit suggestion to use types | ## Code Before:
import os
import subprocess
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
yield {"pub_key": str(pub_key), "priv_key": str(priv_key)}
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
class TestSSHShell:
def test_ssh_shell_key_gen(self, keys):
"""
Test ssh key_gen
"""
shell.gen_key(keys["priv_key"])
for fp in keys.keys():
assert os.path.exists(keys[fp])
# verify there is not a passphrase set on key
ret = subprocess.check_output(
["ssh-keygen", "-f", keys["priv_key"], "-y"], timeout=30,
)
assert ret.decode().startswith("ssh-rsa")
## Instruction:
Use commit suggestion to use types
## Code After:
import subprocess
import types
import pytest
import salt.client.ssh.shell as shell
@pytest.fixture
def keys(tmp_path):
pub_key = tmp_path / "ssh" / "testkey.pub"
priv_key = tmp_path / "ssh" / "testkey"
return types.SimpleNamespace(pub_key=pub_key, priv_key=priv_key)
@pytest.mark.skip_on_windows(reason="Windows does not support salt-ssh")
@pytest.mark.skip_if_binaries_missing("ssh", "ssh-keygen", check_all=True)
def test_ssh_shell_key_gen(keys):
"""
Test ssh key_gen
"""
shell.gen_key(str(keys.priv_key))
assert keys.priv_key.exists()
assert keys.pub_key.exists()
# verify there is not a passphrase set on key
ret = subprocess.check_output(
["ssh-keygen", "-f", str(keys.priv_key), "-y"], timeout=30,
)
assert ret.decode().startswith("ssh-rsa")
|
09d780474d00f3a8f4c2295154d74dae2023c1d3 | samples/storage_sample/storage/__init__.py | samples/storage_sample/storage/__init__.py | """Common imports for generated storage client library."""
# pylint:disable=wildcard-import
import pkgutil
from apitools.base.py import *
from storage_v1 import *
from storage_v1_client import *
from storage_v1_messages import *
__path__ = pkgutil.extend_path(__path__, __name__)
| """Common imports for generated storage client library."""
# pylint:disable=wildcard-import
import pkgutil
from apitools.base.py import *
from storage_v1_client import *
from storage_v1_messages import *
__path__ = pkgutil.extend_path(__path__, __name__)
| Drop the CLI from the sample storage client imports. | Drop the CLI from the sample storage client imports.
| Python | apache-2.0 | cherba/apitools,craigcitro/apitools,b-daniels/apitools,betamos/apitools,kevinli7/apitools,houglum/apitools,pcostell/apitools,thobrla/apitools,google/apitools | """Common imports for generated storage client library."""
# pylint:disable=wildcard-import
import pkgutil
from apitools.base.py import *
- from storage_v1 import *
from storage_v1_client import *
from storage_v1_messages import *
__path__ = pkgutil.extend_path(__path__, __name__)
| Drop the CLI from the sample storage client imports. | ## Code Before:
"""Common imports for generated storage client library."""
# pylint:disable=wildcard-import
import pkgutil
from apitools.base.py import *
from storage_v1 import *
from storage_v1_client import *
from storage_v1_messages import *
__path__ = pkgutil.extend_path(__path__, __name__)
## Instruction:
Drop the CLI from the sample storage client imports.
## Code After:
"""Common imports for generated storage client library."""
# pylint:disable=wildcard-import
import pkgutil
from apitools.base.py import *
from storage_v1_client import *
from storage_v1_messages import *
__path__ = pkgutil.extend_path(__path__, __name__)
|
08d838e87bd92dacbbbfe31b19c628b9d3b271a8 | src/plone.example/plone/example/todo.py | src/plone.example/plone/example/todo.py | from plone.dexterity.interfaces import IDexterityContent
from plone.dexterity.interfaces import IFormFieldProvider
from plone.server.api.service import Service
from plone.supermodel import model
from zope import schema
from zope.component import adapter
from zope.dublincore.annotatableadapter import ZDCAnnotatableAdapter
from zope.dublincore.interfaces import IWriteZopeDublinCore
from zope.interface import provider
class ITodo(model.Schema):
title = schema.TextLine(
title=u"Title",
required=False,
description=u"It's a title",
)
done = schema.Bool(
title=u"Done",
required=False,
description=u"Has the task been completed?",
)
class View(Service):
def __init__(self, context, request):
self.context = context
self.request = request
async def __call__(self):
return {
'context': str(self.context),
'portal_type': self.context.portal_type,
}
@provider(IFormFieldProvider)
class IDublinCore(IWriteZopeDublinCore):
""" We basically just want the IFormFieldProvider interface applied
There's probably a zcml way of doing this. """
@adapter(IDexterityContent)
class DublinCore(ZDCAnnotatableAdapter):
pass
| from plone.dexterity.interfaces import IDexterityContent
from plone.dexterity.interfaces import IFormFieldProvider
from plone.server.api.service import Service
from plone.supermodel import model
from zope import schema
from zope.component import adapter
from zope.dublincore.annotatableadapter import ZDCAnnotatableAdapter
from zope.dublincore.interfaces import IWriteZopeDublinCore
from zope.interface import provider
class ITodo(model.Schema):
title = schema.TextLine(
title=u"Title",
required=False,
description=u"It's a title",
default=u''
)
done = schema.Bool(
title=u"Done",
required=False,
description=u"Has the task been completed?",
default=False
)
class View(Service):
def __init__(self, context, request):
self.context = context
self.request = request
async def __call__(self):
return {
'context': str(self.context),
'portal_type': self.context.portal_type,
}
@provider(IFormFieldProvider)
class IDublinCore(IWriteZopeDublinCore):
""" We basically just want the IFormFieldProvider interface applied
There's probably a zcml way of doing this. """
@adapter(IDexterityContent)
class DublinCore(ZDCAnnotatableAdapter):
pass
| Set default values for fields | Set default values for fields
| Python | bsd-2-clause | plone/plone.server,plone/plone.server | from plone.dexterity.interfaces import IDexterityContent
from plone.dexterity.interfaces import IFormFieldProvider
from plone.server.api.service import Service
from plone.supermodel import model
from zope import schema
from zope.component import adapter
from zope.dublincore.annotatableadapter import ZDCAnnotatableAdapter
from zope.dublincore.interfaces import IWriteZopeDublinCore
from zope.interface import provider
class ITodo(model.Schema):
title = schema.TextLine(
title=u"Title",
required=False,
description=u"It's a title",
+ default=u''
)
done = schema.Bool(
title=u"Done",
required=False,
description=u"Has the task been completed?",
+ default=False
)
class View(Service):
def __init__(self, context, request):
self.context = context
self.request = request
async def __call__(self):
return {
'context': str(self.context),
'portal_type': self.context.portal_type,
}
@provider(IFormFieldProvider)
class IDublinCore(IWriteZopeDublinCore):
""" We basically just want the IFormFieldProvider interface applied
There's probably a zcml way of doing this. """
@adapter(IDexterityContent)
class DublinCore(ZDCAnnotatableAdapter):
pass
| Set default values for fields | ## Code Before:
from plone.dexterity.interfaces import IDexterityContent
from plone.dexterity.interfaces import IFormFieldProvider
from plone.server.api.service import Service
from plone.supermodel import model
from zope import schema
from zope.component import adapter
from zope.dublincore.annotatableadapter import ZDCAnnotatableAdapter
from zope.dublincore.interfaces import IWriteZopeDublinCore
from zope.interface import provider
class ITodo(model.Schema):
title = schema.TextLine(
title=u"Title",
required=False,
description=u"It's a title",
)
done = schema.Bool(
title=u"Done",
required=False,
description=u"Has the task been completed?",
)
class View(Service):
def __init__(self, context, request):
self.context = context
self.request = request
async def __call__(self):
return {
'context': str(self.context),
'portal_type': self.context.portal_type,
}
@provider(IFormFieldProvider)
class IDublinCore(IWriteZopeDublinCore):
""" We basically just want the IFormFieldProvider interface applied
There's probably a zcml way of doing this. """
@adapter(IDexterityContent)
class DublinCore(ZDCAnnotatableAdapter):
pass
## Instruction:
Set default values for fields
## Code After:
from plone.dexterity.interfaces import IDexterityContent
from plone.dexterity.interfaces import IFormFieldProvider
from plone.server.api.service import Service
from plone.supermodel import model
from zope import schema
from zope.component import adapter
from zope.dublincore.annotatableadapter import ZDCAnnotatableAdapter
from zope.dublincore.interfaces import IWriteZopeDublinCore
from zope.interface import provider
class ITodo(model.Schema):
title = schema.TextLine(
title=u"Title",
required=False,
description=u"It's a title",
default=u''
)
done = schema.Bool(
title=u"Done",
required=False,
description=u"Has the task been completed?",
default=False
)
class View(Service):
def __init__(self, context, request):
self.context = context
self.request = request
async def __call__(self):
return {
'context': str(self.context),
'portal_type': self.context.portal_type,
}
@provider(IFormFieldProvider)
class IDublinCore(IWriteZopeDublinCore):
""" We basically just want the IFormFieldProvider interface applied
There's probably a zcml way of doing this. """
@adapter(IDexterityContent)
class DublinCore(ZDCAnnotatableAdapter):
pass
|
29a964a64230e26fca550e81a1ecba3dd782dfb1 | python/vtd.py | python/vtd.py | import libvtd.trusted_system
def UpdateTrustedSystem(file_name):
"""Make sure the TrustedSystem object is up to date."""
global my_system
my_system = libvtd.trusted_system.TrustedSystem()
my_system.AddFile(file_name)
| import libvtd.trusted_system
def UpdateTrustedSystem(file_name):
"""Make sure the TrustedSystem object is up to date."""
global my_system
if 'my_system' not in globals():
my_system = libvtd.trusted_system.TrustedSystem()
my_system.AddFile(file_name)
my_system.Refresh()
| Refresh system instead of clobbering it | Refresh system instead of clobbering it
Otherwise, if we set the Contexts, they'll be gone before we can request the
NextActions!
| Python | apache-2.0 | chiphogg/vim-vtd | import libvtd.trusted_system
def UpdateTrustedSystem(file_name):
"""Make sure the TrustedSystem object is up to date."""
global my_system
+ if 'my_system' not in globals():
- my_system = libvtd.trusted_system.TrustedSystem()
+ my_system = libvtd.trusted_system.TrustedSystem()
- my_system.AddFile(file_name)
+ my_system.AddFile(file_name)
+ my_system.Refresh()
| Refresh system instead of clobbering it | ## Code Before:
import libvtd.trusted_system
def UpdateTrustedSystem(file_name):
"""Make sure the TrustedSystem object is up to date."""
global my_system
my_system = libvtd.trusted_system.TrustedSystem()
my_system.AddFile(file_name)
## Instruction:
Refresh system instead of clobbering it
## Code After:
import libvtd.trusted_system
def UpdateTrustedSystem(file_name):
"""Make sure the TrustedSystem object is up to date."""
global my_system
if 'my_system' not in globals():
my_system = libvtd.trusted_system.TrustedSystem()
my_system.AddFile(file_name)
my_system.Refresh()
|
b532ffff18e95b6014921d88b6df075e8ac2c4ec | problib/example1/__init__.py | problib/example1/__init__.py | from sympy import symbols, cos, sin, latex
from mathdeck import rand, answer
metadata = {
'author': 'Bob Hope',
'institution': 'University of Missouri',
'subject': 'algebra',
'minor subject': 'polynomial equations',
'tags': ['simplify','roots','intervals']
}
r = rand.Random()
# choose three random integers between 0 and 10.
root1 = r.randint(0,10)
root2 = r.randint(0,10)
root3 = r.randint(0,10)
#
# # specify our variables
x = symbols('x')
p = ((x-root1)*(x-root2)).expand(basic=True)
template_variables = {
'p': latex(p),
}
a1 = answer.Answer()
a1.value = cos(x)**2-sin(x)**2
a1.type = 'function'
a1.variables = ['x']
a1.domain = 'R'
a2 = answer.Answer()
a2.value = 'x+1'
a2.type = "function"
a2.variables = ['x','y']
answers = {
'ans1': a1,
'ans2': a2
}
| from sympy import symbols, cos, sin, latex
from mathdeck import rand, answer
metadata = {
'author': 'Bob Hope',
'institution': 'University of Missouri',
'subject': 'algebra',
'minor subject': 'polynomial equations',
'tags': ['simplify','roots','intervals']
}
r = rand.Random()
# choose three random integers between 0 and 10.
root1 = r.randint(0,10)
root2 = r.randint(0,10)
root3 = r.randint(0,10)
#
# # specify our variables
x = symbols('x')
p = ((x-root1)*(x-root2)).expand(basic=True)
func = cos(x)**2-sin(x)**2
a1 = answer.Answer(
value=func,
type='function',
vars=['x'])
a2 = answer.Answer(value='x+1',type='function',vars=['x'])
answers = {
'ans1': a1,
'ans2': a2
}
template_variables = {
'p': latex(p),
}
| Update mathdeck problib for new Answer refactoring | Update mathdeck problib for new Answer refactoring
| Python | apache-2.0 | patrickspencer/mathdeck,patrickspencer/mathdeck | from sympy import symbols, cos, sin, latex
from mathdeck import rand, answer
metadata = {
'author': 'Bob Hope',
'institution': 'University of Missouri',
'subject': 'algebra',
'minor subject': 'polynomial equations',
'tags': ['simplify','roots','intervals']
}
r = rand.Random()
# choose three random integers between 0 and 10.
root1 = r.randint(0,10)
root2 = r.randint(0,10)
root3 = r.randint(0,10)
#
# # specify our variables
x = symbols('x')
p = ((x-root1)*(x-root2)).expand(basic=True)
+ func = cos(x)**2-sin(x)**2
- template_variables = {
- 'p': latex(p),
- }
- a1 = answer.Answer()
+ a1 = answer.Answer(
+ value=func,
+ type='function',
+ vars=['x'])
+ a2 = answer.Answer(value='x+1',type='function',vars=['x'])
- a1.value = cos(x)**2-sin(x)**2
- a1.type = 'function'
- a1.variables = ['x']
- a1.domain = 'R'
-
- a2 = answer.Answer()
- a2.value = 'x+1'
- a2.type = "function"
- a2.variables = ['x','y']
answers = {
'ans1': a1,
'ans2': a2
}
+ template_variables = {
+ 'p': latex(p),
+ }
+ | Update mathdeck problib for new Answer refactoring | ## Code Before:
from sympy import symbols, cos, sin, latex
from mathdeck import rand, answer
metadata = {
'author': 'Bob Hope',
'institution': 'University of Missouri',
'subject': 'algebra',
'minor subject': 'polynomial equations',
'tags': ['simplify','roots','intervals']
}
r = rand.Random()
# choose three random integers between 0 and 10.
root1 = r.randint(0,10)
root2 = r.randint(0,10)
root3 = r.randint(0,10)
#
# # specify our variables
x = symbols('x')
p = ((x-root1)*(x-root2)).expand(basic=True)
template_variables = {
'p': latex(p),
}
a1 = answer.Answer()
a1.value = cos(x)**2-sin(x)**2
a1.type = 'function'
a1.variables = ['x']
a1.domain = 'R'
a2 = answer.Answer()
a2.value = 'x+1'
a2.type = "function"
a2.variables = ['x','y']
answers = {
'ans1': a1,
'ans2': a2
}
## Instruction:
Update mathdeck problib for new Answer refactoring
## Code After:
from sympy import symbols, cos, sin, latex
from mathdeck import rand, answer
metadata = {
'author': 'Bob Hope',
'institution': 'University of Missouri',
'subject': 'algebra',
'minor subject': 'polynomial equations',
'tags': ['simplify','roots','intervals']
}
r = rand.Random()
# choose three random integers between 0 and 10.
root1 = r.randint(0,10)
root2 = r.randint(0,10)
root3 = r.randint(0,10)
#
# # specify our variables
x = symbols('x')
p = ((x-root1)*(x-root2)).expand(basic=True)
func = cos(x)**2-sin(x)**2
a1 = answer.Answer(
value=func,
type='function',
vars=['x'])
a2 = answer.Answer(value='x+1',type='function',vars=['x'])
answers = {
'ans1': a1,
'ans2': a2
}
template_variables = {
'p': latex(p),
}
|
3a9568b4d4de969b1e2031e8d2d3cdd7bd56824f | zerver/migrations/0237_rename_zulip_realm_to_zulipinternal.py | zerver/migrations/0237_rename_zulip_realm_to_zulipinternal.py | from django.conf import settings
from django.db import migrations
from django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor
from django.db.migrations.state import StateApps
def rename_zulip_realm_to_zulipinternal(apps: StateApps, schema_editor: DatabaseSchemaEditor) -> None:
if not settings.PRODUCTION:
return
Realm = apps.get_model('zerver', 'Realm')
UserProfile = apps.get_model('zerver', 'UserProfile')
if Realm.objects.count() == 0:
# Database not yet populated, do nothing:
return
if Realm.objects.filter(string_id="zulipinternal").exists():
return
internal_realm = Realm.objects.get(string_id="zulip")
# For safety, as a sanity check, verify that "internal_realm" is indeed the realm for system bots:
welcome_bot = UserProfile.objects.get(email="welcome-bot@zulip.com")
assert welcome_bot.realm.id == internal_realm.id
internal_realm.string_id = "zulipinternal"
internal_realm.name = "System use only"
internal_realm.save()
class Migration(migrations.Migration):
dependencies = [
('zerver', '0236_remove_illegal_characters_email_full'),
]
operations = [
migrations.RunPython(rename_zulip_realm_to_zulipinternal)
]
| from django.conf import settings
from django.db import migrations
from django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor
from django.db.migrations.state import StateApps
def rename_zulip_realm_to_zulipinternal(apps: StateApps, schema_editor: DatabaseSchemaEditor) -> None:
if not settings.PRODUCTION:
return
Realm = apps.get_model('zerver', 'Realm')
UserProfile = apps.get_model('zerver', 'UserProfile')
if Realm.objects.count() == 0:
# Database not yet populated, do nothing:
return
if Realm.objects.filter(string_id="zulipinternal").exists():
return
if not Realm.objects.filter(string_id="zulip").exists():
# If the user renamed the `zulip` system bot realm (or deleted
# it), there's nothing for us to do.
return
internal_realm = Realm.objects.get(string_id="zulip")
# For safety, as a sanity check, verify that "internal_realm" is indeed the realm for system bots:
welcome_bot = UserProfile.objects.get(email="welcome-bot@zulip.com")
assert welcome_bot.realm.id == internal_realm.id
internal_realm.string_id = "zulipinternal"
internal_realm.name = "System use only"
internal_realm.save()
class Migration(migrations.Migration):
dependencies = [
('zerver', '0236_remove_illegal_characters_email_full'),
]
operations = [
migrations.RunPython(rename_zulip_realm_to_zulipinternal)
]
| Fix zulipinternal migration corner case. | migrations: Fix zulipinternal migration corner case.
It's theoretically possible to have configured a Zulip server where
the system bots live in the same realm as normal users (and may have
in fact been the default in early Zulip releases? Unclear.). We
should handle these without the migration intended to clean up naming
for the system bot realm crashing.
Fixes #13660.
| Python | apache-2.0 | brainwane/zulip,andersk/zulip,hackerkid/zulip,synicalsyntax/zulip,andersk/zulip,hackerkid/zulip,punchagan/zulip,shubhamdhama/zulip,showell/zulip,zulip/zulip,synicalsyntax/zulip,kou/zulip,rht/zulip,showell/zulip,brainwane/zulip,punchagan/zulip,shubhamdhama/zulip,hackerkid/zulip,andersk/zulip,kou/zulip,brainwane/zulip,brainwane/zulip,andersk/zulip,showell/zulip,shubhamdhama/zulip,kou/zulip,punchagan/zulip,zulip/zulip,kou/zulip,showell/zulip,brainwane/zulip,eeshangarg/zulip,timabbott/zulip,eeshangarg/zulip,hackerkid/zulip,shubhamdhama/zulip,zulip/zulip,timabbott/zulip,brainwane/zulip,eeshangarg/zulip,rht/zulip,zulip/zulip,timabbott/zulip,rht/zulip,synicalsyntax/zulip,andersk/zulip,showell/zulip,shubhamdhama/zulip,rht/zulip,hackerkid/zulip,rht/zulip,hackerkid/zulip,eeshangarg/zulip,kou/zulip,synicalsyntax/zulip,brainwane/zulip,shubhamdhama/zulip,eeshangarg/zulip,andersk/zulip,rht/zulip,timabbott/zulip,punchagan/zulip,showell/zulip,eeshangarg/zulip,synicalsyntax/zulip,punchagan/zulip,timabbott/zulip,zulip/zulip,andersk/zulip,punchagan/zulip,zulip/zulip,hackerkid/zulip,timabbott/zulip,synicalsyntax/zulip,zulip/zulip,showell/zulip,eeshangarg/zulip,kou/zulip,punchagan/zulip,synicalsyntax/zulip,rht/zulip,kou/zulip,timabbott/zulip,shubhamdhama/zulip | from django.conf import settings
from django.db import migrations
from django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor
from django.db.migrations.state import StateApps
def rename_zulip_realm_to_zulipinternal(apps: StateApps, schema_editor: DatabaseSchemaEditor) -> None:
if not settings.PRODUCTION:
return
Realm = apps.get_model('zerver', 'Realm')
UserProfile = apps.get_model('zerver', 'UserProfile')
if Realm.objects.count() == 0:
# Database not yet populated, do nothing:
return
if Realm.objects.filter(string_id="zulipinternal").exists():
+ return
+ if not Realm.objects.filter(string_id="zulip").exists():
+ # If the user renamed the `zulip` system bot realm (or deleted
+ # it), there's nothing for us to do.
return
internal_realm = Realm.objects.get(string_id="zulip")
# For safety, as a sanity check, verify that "internal_realm" is indeed the realm for system bots:
welcome_bot = UserProfile.objects.get(email="welcome-bot@zulip.com")
assert welcome_bot.realm.id == internal_realm.id
internal_realm.string_id = "zulipinternal"
internal_realm.name = "System use only"
internal_realm.save()
class Migration(migrations.Migration):
dependencies = [
('zerver', '0236_remove_illegal_characters_email_full'),
]
operations = [
migrations.RunPython(rename_zulip_realm_to_zulipinternal)
]
| Fix zulipinternal migration corner case. | ## Code Before:
from django.conf import settings
from django.db import migrations
from django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor
from django.db.migrations.state import StateApps
def rename_zulip_realm_to_zulipinternal(apps: StateApps, schema_editor: DatabaseSchemaEditor) -> None:
if not settings.PRODUCTION:
return
Realm = apps.get_model('zerver', 'Realm')
UserProfile = apps.get_model('zerver', 'UserProfile')
if Realm.objects.count() == 0:
# Database not yet populated, do nothing:
return
if Realm.objects.filter(string_id="zulipinternal").exists():
return
internal_realm = Realm.objects.get(string_id="zulip")
# For safety, as a sanity check, verify that "internal_realm" is indeed the realm for system bots:
welcome_bot = UserProfile.objects.get(email="welcome-bot@zulip.com")
assert welcome_bot.realm.id == internal_realm.id
internal_realm.string_id = "zulipinternal"
internal_realm.name = "System use only"
internal_realm.save()
class Migration(migrations.Migration):
dependencies = [
('zerver', '0236_remove_illegal_characters_email_full'),
]
operations = [
migrations.RunPython(rename_zulip_realm_to_zulipinternal)
]
## Instruction:
Fix zulipinternal migration corner case.
## Code After:
from django.conf import settings
from django.db import migrations
from django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor
from django.db.migrations.state import StateApps
def rename_zulip_realm_to_zulipinternal(apps: StateApps, schema_editor: DatabaseSchemaEditor) -> None:
if not settings.PRODUCTION:
return
Realm = apps.get_model('zerver', 'Realm')
UserProfile = apps.get_model('zerver', 'UserProfile')
if Realm.objects.count() == 0:
# Database not yet populated, do nothing:
return
if Realm.objects.filter(string_id="zulipinternal").exists():
return
if not Realm.objects.filter(string_id="zulip").exists():
# If the user renamed the `zulip` system bot realm (or deleted
# it), there's nothing for us to do.
return
internal_realm = Realm.objects.get(string_id="zulip")
# For safety, as a sanity check, verify that "internal_realm" is indeed the realm for system bots:
welcome_bot = UserProfile.objects.get(email="welcome-bot@zulip.com")
assert welcome_bot.realm.id == internal_realm.id
internal_realm.string_id = "zulipinternal"
internal_realm.name = "System use only"
internal_realm.save()
class Migration(migrations.Migration):
dependencies = [
('zerver', '0236_remove_illegal_characters_email_full'),
]
operations = [
migrations.RunPython(rename_zulip_realm_to_zulipinternal)
]
|
97e3b202bbe6726a4056facb8b4690b0710029a9 | handroll/tests/test_site.py | handroll/tests/test_site.py |
import os
import tempfile
from handroll.site import Site
from handroll.tests import TestCase
class TestSite(TestCase):
def test_finds_valid_site_root_from_templates(self):
original = os.getcwd()
valid_site = tempfile.mkdtemp()
open(os.path.join(valid_site, 'template.html'), 'w').close()
os.chdir(valid_site)
site = Site()
self.assertEqual(valid_site, site.path)
os.chdir(original)
def test_finds_valid_site_root_from_conf(self):
original = os.getcwd()
valid_site = tempfile.mkdtemp()
open(os.path.join(valid_site, Site.CONFIG), 'w').close()
os.chdir(valid_site)
site = Site()
self.assertEqual(valid_site, site.path)
os.chdir(original)
def test_site_has_absolute_path(self):
original = os.getcwd()
tempdir = tempfile.mkdtemp()
site_path = os.path.join(tempdir, 'site')
os.mkdir(site_path)
os.chdir(tempdir)
site = Site('site')
self.assertEqual(site_path, site.path)
os.chdir(original)
|
import os
import tempfile
from handroll.site import Site
from handroll.tests import TestCase
class TestSite(TestCase):
def test_finds_valid_site_root_from_templates(self):
original = os.getcwd()
valid_site = os.path.realpath(tempfile.mkdtemp())
open(os.path.join(valid_site, 'template.html'), 'w').close()
os.chdir(valid_site)
site = Site()
self.assertEqual(valid_site, site.path)
os.chdir(original)
def test_finds_valid_site_root_from_conf(self):
original = os.getcwd()
valid_site = os.path.realpath(tempfile.mkdtemp())
open(os.path.join(valid_site, Site.CONFIG), 'w').close()
os.chdir(valid_site)
site = Site()
self.assertEqual(valid_site, site.path)
os.chdir(original)
def test_site_has_absolute_path(self):
original = os.getcwd()
tempdir = os.path.realpath(tempfile.mkdtemp())
site_path = os.path.join(tempdir, 'site')
os.mkdir(site_path)
os.chdir(tempdir)
site = Site('site')
self.assertEqual(site_path, site.path)
os.chdir(original)
| Use a real path when testing sites. | Use a real path when testing sites.
Mac OS X returns link paths when calling `mkdtemp`. Calling realpath
allows the site path comparison to be consistent across platforms.
| Python | bsd-2-clause | handroll/handroll |
import os
import tempfile
from handroll.site import Site
from handroll.tests import TestCase
class TestSite(TestCase):
def test_finds_valid_site_root_from_templates(self):
original = os.getcwd()
- valid_site = tempfile.mkdtemp()
+ valid_site = os.path.realpath(tempfile.mkdtemp())
open(os.path.join(valid_site, 'template.html'), 'w').close()
os.chdir(valid_site)
site = Site()
self.assertEqual(valid_site, site.path)
os.chdir(original)
def test_finds_valid_site_root_from_conf(self):
original = os.getcwd()
- valid_site = tempfile.mkdtemp()
+ valid_site = os.path.realpath(tempfile.mkdtemp())
open(os.path.join(valid_site, Site.CONFIG), 'w').close()
os.chdir(valid_site)
site = Site()
self.assertEqual(valid_site, site.path)
os.chdir(original)
def test_site_has_absolute_path(self):
original = os.getcwd()
- tempdir = tempfile.mkdtemp()
+ tempdir = os.path.realpath(tempfile.mkdtemp())
site_path = os.path.join(tempdir, 'site')
os.mkdir(site_path)
os.chdir(tempdir)
site = Site('site')
self.assertEqual(site_path, site.path)
os.chdir(original)
| Use a real path when testing sites. | ## Code Before:
import os
import tempfile
from handroll.site import Site
from handroll.tests import TestCase
class TestSite(TestCase):
def test_finds_valid_site_root_from_templates(self):
original = os.getcwd()
valid_site = tempfile.mkdtemp()
open(os.path.join(valid_site, 'template.html'), 'w').close()
os.chdir(valid_site)
site = Site()
self.assertEqual(valid_site, site.path)
os.chdir(original)
def test_finds_valid_site_root_from_conf(self):
original = os.getcwd()
valid_site = tempfile.mkdtemp()
open(os.path.join(valid_site, Site.CONFIG), 'w').close()
os.chdir(valid_site)
site = Site()
self.assertEqual(valid_site, site.path)
os.chdir(original)
def test_site_has_absolute_path(self):
original = os.getcwd()
tempdir = tempfile.mkdtemp()
site_path = os.path.join(tempdir, 'site')
os.mkdir(site_path)
os.chdir(tempdir)
site = Site('site')
self.assertEqual(site_path, site.path)
os.chdir(original)
## Instruction:
Use a real path when testing sites.
## Code After:
import os
import tempfile
from handroll.site import Site
from handroll.tests import TestCase
class TestSite(TestCase):
def test_finds_valid_site_root_from_templates(self):
original = os.getcwd()
valid_site = os.path.realpath(tempfile.mkdtemp())
open(os.path.join(valid_site, 'template.html'), 'w').close()
os.chdir(valid_site)
site = Site()
self.assertEqual(valid_site, site.path)
os.chdir(original)
def test_finds_valid_site_root_from_conf(self):
original = os.getcwd()
valid_site = os.path.realpath(tempfile.mkdtemp())
open(os.path.join(valid_site, Site.CONFIG), 'w').close()
os.chdir(valid_site)
site = Site()
self.assertEqual(valid_site, site.path)
os.chdir(original)
def test_site_has_absolute_path(self):
original = os.getcwd()
tempdir = os.path.realpath(tempfile.mkdtemp())
site_path = os.path.join(tempdir, 'site')
os.mkdir(site_path)
os.chdir(tempdir)
site = Site('site')
self.assertEqual(site_path, site.path)
os.chdir(original)
|
7e44a8bd38105144111624710819a1ee54891222 | campos_checkin/__openerp__.py | campos_checkin/__openerp__.py |
{
'name': 'Campos Checkin',
'description': """
CampOS Check In functionality""",
'version': '8.0.1.0.0',
'license': 'AGPL-3',
'author': 'Stein & Gabelgaard ApS',
'website': 'www.steingabelgaard.dk',
'depends': [
'campos_jobber_final',
'campos_transportation',
'campos_crewnet',
'web_ir_actions_act_window_message',
#'web_tree_dynamic_colored_field',
],
'data': [
'wizards/campos_checkin_grp_wiz.xml',
'views/event_registration.xml',
'wizards/campos_checkin_wiz.xml',
'security/campos_checkin.xml',
'views/campos_event_participant.xml',
'views/campos_mat_report.xml',
],
'demo': [
],
}
|
{
'name': 'Campos Checkin',
'description': """
CampOS Check In functionality""",
'version': '8.0.1.0.0',
'license': 'AGPL-3',
'author': 'Stein & Gabelgaard ApS',
'website': 'www.steingabelgaard.dk',
'depends': [
'campos_jobber_final',
'campos_transportation',
'campos_crewnet',
'web_ir_actions_act_window_message',
#'web_tree_dynamic_colored_field',
],
'data': [
'wizards/campos_checkin_wiz.xml',
'security/campos_checkin.xml',
'views/campos_event_participant.xml',
'views/campos_mat_report.xml',
'wizards/campos_checkin_grp_wiz.xml',
'views/event_registration.xml',
],
'demo': [
],
}
| Fix order for menu ref | Fix order for menu ref | Python | agpl-3.0 | sl2017/campos |
{
'name': 'Campos Checkin',
'description': """
CampOS Check In functionality""",
'version': '8.0.1.0.0',
'license': 'AGPL-3',
'author': 'Stein & Gabelgaard ApS',
'website': 'www.steingabelgaard.dk',
'depends': [
'campos_jobber_final',
'campos_transportation',
'campos_crewnet',
'web_ir_actions_act_window_message',
#'web_tree_dynamic_colored_field',
],
'data': [
- 'wizards/campos_checkin_grp_wiz.xml',
- 'views/event_registration.xml',
'wizards/campos_checkin_wiz.xml',
'security/campos_checkin.xml',
'views/campos_event_participant.xml',
'views/campos_mat_report.xml',
+ 'wizards/campos_checkin_grp_wiz.xml',
+ 'views/event_registration.xml',
],
'demo': [
],
}
| Fix order for menu ref | ## Code Before:
{
'name': 'Campos Checkin',
'description': """
CampOS Check In functionality""",
'version': '8.0.1.0.0',
'license': 'AGPL-3',
'author': 'Stein & Gabelgaard ApS',
'website': 'www.steingabelgaard.dk',
'depends': [
'campos_jobber_final',
'campos_transportation',
'campos_crewnet',
'web_ir_actions_act_window_message',
#'web_tree_dynamic_colored_field',
],
'data': [
'wizards/campos_checkin_grp_wiz.xml',
'views/event_registration.xml',
'wizards/campos_checkin_wiz.xml',
'security/campos_checkin.xml',
'views/campos_event_participant.xml',
'views/campos_mat_report.xml',
],
'demo': [
],
}
## Instruction:
Fix order for menu ref
## Code After:
{
'name': 'Campos Checkin',
'description': """
CampOS Check In functionality""",
'version': '8.0.1.0.0',
'license': 'AGPL-3',
'author': 'Stein & Gabelgaard ApS',
'website': 'www.steingabelgaard.dk',
'depends': [
'campos_jobber_final',
'campos_transportation',
'campos_crewnet',
'web_ir_actions_act_window_message',
#'web_tree_dynamic_colored_field',
],
'data': [
'wizards/campos_checkin_wiz.xml',
'security/campos_checkin.xml',
'views/campos_event_participant.xml',
'views/campos_mat_report.xml',
'wizards/campos_checkin_grp_wiz.xml',
'views/event_registration.xml',
],
'demo': [
],
}
|
3cd99c23099a625da711e3ac458a46a7b364d83c | hystrix/pool.py | hystrix/pool.py | from __future__ import absolute_import
from concurrent.futures import ThreadPoolExecutor
import logging
import six
log = logging.getLogger(__name__)
class PoolMetaclass(type):
__instances__ = dict()
__blacklist__ = ('Pool', 'PoolMetaclass')
def __new__(cls, name, bases, attrs):
if name in cls.__blacklist__:
return super(PoolMetaclass, cls).__new__(cls, name,
bases, attrs)
pool_key = attrs.get('pool_key') or '{}Pool'.format(name)
new_class = super(PoolMetaclass, cls).__new__(cls, pool_key,
bases, attrs)
setattr(new_class, 'pool_key', pool_key)
if pool_key not in cls.__instances__:
cls.__instances__[pool_key] = new_class
return cls.__instances__[pool_key]
class Pool(six.with_metaclass(PoolMetaclass, ThreadPoolExecutor)):
pool_key = None
def __init__(self, pool_key=None, max_workers=5):
super(Pool, self).__init__(max_workers)
| from __future__ import absolute_import
from concurrent.futures import ProcessPoolExecutor
import logging
import six
log = logging.getLogger(__name__)
class PoolMetaclass(type):
__instances__ = dict()
__blacklist__ = ('Pool', 'PoolMetaclass')
def __new__(cls, name, bases, attrs):
if name in cls.__blacklist__:
return super(PoolMetaclass, cls).__new__(cls, name,
bases, attrs)
pool_key = attrs.get('pool_key') or '{}Pool'.format(name)
new_class = super(PoolMetaclass, cls).__new__(cls, pool_key,
bases, attrs)
setattr(new_class, 'pool_key', pool_key)
if pool_key not in cls.__instances__:
cls.__instances__[pool_key] = new_class
return cls.__instances__[pool_key]
class Pool(six.with_metaclass(PoolMetaclass, ProcessPoolExecutor)):
pool_key = None
def __init__(self, pool_key=None, max_workers=5):
super(Pool, self).__init__(max_workers)
| Change Pool to use ProcessPoolExecutor | Change Pool to use ProcessPoolExecutor
| Python | apache-2.0 | wiliamsouza/hystrix-py,wiliamsouza/hystrix-py | from __future__ import absolute_import
- from concurrent.futures import ThreadPoolExecutor
+ from concurrent.futures import ProcessPoolExecutor
import logging
import six
log = logging.getLogger(__name__)
class PoolMetaclass(type):
__instances__ = dict()
__blacklist__ = ('Pool', 'PoolMetaclass')
def __new__(cls, name, bases, attrs):
if name in cls.__blacklist__:
return super(PoolMetaclass, cls).__new__(cls, name,
bases, attrs)
pool_key = attrs.get('pool_key') or '{}Pool'.format(name)
new_class = super(PoolMetaclass, cls).__new__(cls, pool_key,
bases, attrs)
setattr(new_class, 'pool_key', pool_key)
if pool_key not in cls.__instances__:
cls.__instances__[pool_key] = new_class
return cls.__instances__[pool_key]
- class Pool(six.with_metaclass(PoolMetaclass, ThreadPoolExecutor)):
+ class Pool(six.with_metaclass(PoolMetaclass, ProcessPoolExecutor)):
pool_key = None
def __init__(self, pool_key=None, max_workers=5):
super(Pool, self).__init__(max_workers)
| Change Pool to use ProcessPoolExecutor | ## Code Before:
from __future__ import absolute_import
from concurrent.futures import ThreadPoolExecutor
import logging
import six
log = logging.getLogger(__name__)
class PoolMetaclass(type):
__instances__ = dict()
__blacklist__ = ('Pool', 'PoolMetaclass')
def __new__(cls, name, bases, attrs):
if name in cls.__blacklist__:
return super(PoolMetaclass, cls).__new__(cls, name,
bases, attrs)
pool_key = attrs.get('pool_key') or '{}Pool'.format(name)
new_class = super(PoolMetaclass, cls).__new__(cls, pool_key,
bases, attrs)
setattr(new_class, 'pool_key', pool_key)
if pool_key not in cls.__instances__:
cls.__instances__[pool_key] = new_class
return cls.__instances__[pool_key]
class Pool(six.with_metaclass(PoolMetaclass, ThreadPoolExecutor)):
pool_key = None
def __init__(self, pool_key=None, max_workers=5):
super(Pool, self).__init__(max_workers)
## Instruction:
Change Pool to use ProcessPoolExecutor
## Code After:
from __future__ import absolute_import
from concurrent.futures import ProcessPoolExecutor
import logging
import six
log = logging.getLogger(__name__)
class PoolMetaclass(type):
__instances__ = dict()
__blacklist__ = ('Pool', 'PoolMetaclass')
def __new__(cls, name, bases, attrs):
if name in cls.__blacklist__:
return super(PoolMetaclass, cls).__new__(cls, name,
bases, attrs)
pool_key = attrs.get('pool_key') or '{}Pool'.format(name)
new_class = super(PoolMetaclass, cls).__new__(cls, pool_key,
bases, attrs)
setattr(new_class, 'pool_key', pool_key)
if pool_key not in cls.__instances__:
cls.__instances__[pool_key] = new_class
return cls.__instances__[pool_key]
class Pool(six.with_metaclass(PoolMetaclass, ProcessPoolExecutor)):
pool_key = None
def __init__(self, pool_key=None, max_workers=5):
super(Pool, self).__init__(max_workers)
|
c52edc120f38acb079fa364cdb684fc2052d4727 | corehq/messaging/smsbackends/trumpia/urls.py | corehq/messaging/smsbackends/trumpia/urls.py | from django.conf.urls import url
from corehq.messaging.smsbackends.trumpia.views import TrumpiaIncomingView
urlpatterns = [
url(r'^sms/(?P<api_key>[\w-]+)/?$', TrumpiaIncomingView.as_view(),
name=TrumpiaIncomingView.urlname),
]
| from django.conf.urls import url
from corehq.apps.hqwebapp.decorators import waf_allow
from corehq.messaging.smsbackends.trumpia.views import TrumpiaIncomingView
urlpatterns = [
url(r'^sms/(?P<api_key>[\w-]+)/?$', waf_allow('XSS_QUERYSTRING')(TrumpiaIncomingView.as_view()),
name=TrumpiaIncomingView.urlname),
]
| Annotate trumpia url to say it allows XML in the querystring | Annotate trumpia url to say it allows XML in the querystring
| Python | bsd-3-clause | dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq | from django.conf.urls import url
+
+ from corehq.apps.hqwebapp.decorators import waf_allow
from corehq.messaging.smsbackends.trumpia.views import TrumpiaIncomingView
urlpatterns = [
- url(r'^sms/(?P<api_key>[\w-]+)/?$', TrumpiaIncomingView.as_view(),
+ url(r'^sms/(?P<api_key>[\w-]+)/?$', waf_allow('XSS_QUERYSTRING')(TrumpiaIncomingView.as_view()),
name=TrumpiaIncomingView.urlname),
]
| Annotate trumpia url to say it allows XML in the querystring | ## Code Before:
from django.conf.urls import url
from corehq.messaging.smsbackends.trumpia.views import TrumpiaIncomingView
urlpatterns = [
url(r'^sms/(?P<api_key>[\w-]+)/?$', TrumpiaIncomingView.as_view(),
name=TrumpiaIncomingView.urlname),
]
## Instruction:
Annotate trumpia url to say it allows XML in the querystring
## Code After:
from django.conf.urls import url
from corehq.apps.hqwebapp.decorators import waf_allow
from corehq.messaging.smsbackends.trumpia.views import TrumpiaIncomingView
urlpatterns = [
url(r'^sms/(?P<api_key>[\w-]+)/?$', waf_allow('XSS_QUERYSTRING')(TrumpiaIncomingView.as_view()),
name=TrumpiaIncomingView.urlname),
]
|
d159f8201b9d9aeafd24f07a9e39855fc537182d | cocoscore/tools/data_tools.py | cocoscore/tools/data_tools.py | import pandas as pd
def load_data_frame(data_frame_path, sort_reindex=False, class_labels=True):
"""
Load a sentence data set as pandas DataFrame from a given path.
:param data_frame_path: the path to load the pandas DataFrame from
:param sort_reindex: if True, the returned data frame will be sorted by PMID and reindex by 0, 1, 2, ...
:param class_labels: if True, the class label is assumed to be present as the last column
:return: a pandas DataFrame loaded from the given path
"""
column_names = ['pmid', 'paragraph', 'sentence', 'entity1', 'entity2', 'sentence_text']
if class_labels:
column_names.append('class')
data_df = pd.read_csv(data_frame_path, sep='\t', header=None, index_col=False,
names=column_names)
if sort_reindex:
data_df.sort_values('pmid', axis=0, inplace=True, kind='mergesort')
data_df.reset_index(inplace=True, drop=True)
assert data_df.isnull().sum().sum() == 0
return data_df
| import pandas as pd
def load_data_frame(data_frame_path, sort_reindex=False, class_labels=True, match_distance=False):
"""
Load a sentence data set as pandas DataFrame from a given path.
:param data_frame_path: the path to load the pandas DataFrame from
:param sort_reindex: if True, the returned data frame will be sorted by PMID and reindex by 0, 1, 2, ...
:param class_labels: if True, the class label is assumed to be present as the second-to-last column
:param match_distance: if True, the distance between the closest match is assumed to be present as the last column
:return: a pandas DataFrame loaded from the given path
"""
column_names = ['pmid', 'paragraph', 'sentence', 'entity1', 'entity2', 'sentence_text']
if class_labels:
column_names.append('class')
if match_distance:
column_names.append('distance')
data_df = pd.read_csv(data_frame_path, sep='\t', header=None, index_col=False,
names=column_names)
if sort_reindex:
data_df.sort_values('pmid', axis=0, inplace=True, kind='mergesort')
data_df.reset_index(inplace=True, drop=True)
assert data_df.isnull().sum().sum() == 0
return data_df
| Add match_distance flag to load_data_frame() | Add match_distance flag to load_data_frame()
| Python | mit | JungeAlexander/cocoscore | import pandas as pd
- def load_data_frame(data_frame_path, sort_reindex=False, class_labels=True):
+ def load_data_frame(data_frame_path, sort_reindex=False, class_labels=True, match_distance=False):
"""
Load a sentence data set as pandas DataFrame from a given path.
:param data_frame_path: the path to load the pandas DataFrame from
:param sort_reindex: if True, the returned data frame will be sorted by PMID and reindex by 0, 1, 2, ...
- :param class_labels: if True, the class label is assumed to be present as the last column
+ :param class_labels: if True, the class label is assumed to be present as the second-to-last column
+ :param match_distance: if True, the distance between the closest match is assumed to be present as the last column
:return: a pandas DataFrame loaded from the given path
"""
column_names = ['pmid', 'paragraph', 'sentence', 'entity1', 'entity2', 'sentence_text']
if class_labels:
column_names.append('class')
+ if match_distance:
+ column_names.append('distance')
data_df = pd.read_csv(data_frame_path, sep='\t', header=None, index_col=False,
names=column_names)
if sort_reindex:
data_df.sort_values('pmid', axis=0, inplace=True, kind='mergesort')
data_df.reset_index(inplace=True, drop=True)
assert data_df.isnull().sum().sum() == 0
return data_df
| Add match_distance flag to load_data_frame() | ## Code Before:
import pandas as pd
def load_data_frame(data_frame_path, sort_reindex=False, class_labels=True):
"""
Load a sentence data set as pandas DataFrame from a given path.
:param data_frame_path: the path to load the pandas DataFrame from
:param sort_reindex: if True, the returned data frame will be sorted by PMID and reindex by 0, 1, 2, ...
:param class_labels: if True, the class label is assumed to be present as the last column
:return: a pandas DataFrame loaded from the given path
"""
column_names = ['pmid', 'paragraph', 'sentence', 'entity1', 'entity2', 'sentence_text']
if class_labels:
column_names.append('class')
data_df = pd.read_csv(data_frame_path, sep='\t', header=None, index_col=False,
names=column_names)
if sort_reindex:
data_df.sort_values('pmid', axis=0, inplace=True, kind='mergesort')
data_df.reset_index(inplace=True, drop=True)
assert data_df.isnull().sum().sum() == 0
return data_df
## Instruction:
Add match_distance flag to load_data_frame()
## Code After:
import pandas as pd
def load_data_frame(data_frame_path, sort_reindex=False, class_labels=True, match_distance=False):
"""
Load a sentence data set as pandas DataFrame from a given path.
:param data_frame_path: the path to load the pandas DataFrame from
:param sort_reindex: if True, the returned data frame will be sorted by PMID and reindex by 0, 1, 2, ...
:param class_labels: if True, the class label is assumed to be present as the second-to-last column
:param match_distance: if True, the distance between the closest match is assumed to be present as the last column
:return: a pandas DataFrame loaded from the given path
"""
column_names = ['pmid', 'paragraph', 'sentence', 'entity1', 'entity2', 'sentence_text']
if class_labels:
column_names.append('class')
if match_distance:
column_names.append('distance')
data_df = pd.read_csv(data_frame_path, sep='\t', header=None, index_col=False,
names=column_names)
if sort_reindex:
data_df.sort_values('pmid', axis=0, inplace=True, kind='mergesort')
data_df.reset_index(inplace=True, drop=True)
assert data_df.isnull().sum().sum() == 0
return data_df
|
920e2fbb7e99c17dbe8d5b71e9c9b26a718ca444 | ideascube/search/apps.py | ideascube/search/apps.py | from django.apps import AppConfig
from django.db.models.signals import pre_migrate, post_migrate
from .utils import create_index_table, reindex_content
def create_index(sender, **kwargs):
if isinstance(sender, SearchConfig):
create_index_table(force=True)
def reindex(sender, **kwargs):
if isinstance(sender, SearchConfig):
reindex_content(force=False)
class SearchConfig(AppConfig):
name = 'ideascube.search'
verbose_name = 'Search'
def ready(self):
pre_migrate.connect(create_index, sender=self)
post_migrate.connect(reindex, sender=self)
| from django.apps import AppConfig
from django.db.models.signals import pre_migrate, post_migrate
from .utils import create_index_table, reindex_content
def create_index(sender, **kwargs):
if (kwargs['using'] == 'transient' and isinstance(sender, SearchConfig)):
create_index_table(force=True)
def reindex(sender, **kwargs):
if (kwargs['using'] == 'transient' and isinstance(sender, SearchConfig)):
reindex_content(force=False)
class SearchConfig(AppConfig):
name = 'ideascube.search'
verbose_name = 'Search'
def ready(self):
pre_migrate.connect(create_index, sender=self)
post_migrate.connect(reindex, sender=self)
| Make (pre|post)_migrate scripts for the index table only if working on 'transient'. | Make (pre|post)_migrate scripts for the index table only if working on 'transient'.
Django run (pre|post)_migrate script once per database.
As we have two databases, the create_index is launch twice with different
kwargs['using'] ('default' and 'transient'). We should try to create
the index table only when we are working on the transient database.
Most of the time, this is not important and create a new index table
twice is not important.
However, if we run tests, the database are configured and migrate
one after the other and the 'transient' database may be miss-configured
at a time. By creating the table only at the right time, we ensure that
everything is properly configured.
| Python | agpl-3.0 | ideascube/ideascube,ideascube/ideascube,ideascube/ideascube,ideascube/ideascube | from django.apps import AppConfig
from django.db.models.signals import pre_migrate, post_migrate
from .utils import create_index_table, reindex_content
def create_index(sender, **kwargs):
- if isinstance(sender, SearchConfig):
+ if (kwargs['using'] == 'transient' and isinstance(sender, SearchConfig)):
create_index_table(force=True)
def reindex(sender, **kwargs):
- if isinstance(sender, SearchConfig):
+ if (kwargs['using'] == 'transient' and isinstance(sender, SearchConfig)):
reindex_content(force=False)
class SearchConfig(AppConfig):
name = 'ideascube.search'
verbose_name = 'Search'
def ready(self):
pre_migrate.connect(create_index, sender=self)
post_migrate.connect(reindex, sender=self)
| Make (pre|post)_migrate scripts for the index table only if working on 'transient'. | ## Code Before:
from django.apps import AppConfig
from django.db.models.signals import pre_migrate, post_migrate
from .utils import create_index_table, reindex_content
def create_index(sender, **kwargs):
if isinstance(sender, SearchConfig):
create_index_table(force=True)
def reindex(sender, **kwargs):
if isinstance(sender, SearchConfig):
reindex_content(force=False)
class SearchConfig(AppConfig):
name = 'ideascube.search'
verbose_name = 'Search'
def ready(self):
pre_migrate.connect(create_index, sender=self)
post_migrate.connect(reindex, sender=self)
## Instruction:
Make (pre|post)_migrate scripts for the index table only if working on 'transient'.
## Code After:
from django.apps import AppConfig
from django.db.models.signals import pre_migrate, post_migrate
from .utils import create_index_table, reindex_content
def create_index(sender, **kwargs):
if (kwargs['using'] == 'transient' and isinstance(sender, SearchConfig)):
create_index_table(force=True)
def reindex(sender, **kwargs):
if (kwargs['using'] == 'transient' and isinstance(sender, SearchConfig)):
reindex_content(force=False)
class SearchConfig(AppConfig):
name = 'ideascube.search'
verbose_name = 'Search'
def ready(self):
pre_migrate.connect(create_index, sender=self)
post_migrate.connect(reindex, sender=self)
|
360efe51bc45f189c235bed6b2b7bfdd4fd1bfbd | flask-restful/api.py | flask-restful/api.py | from flask import Flask, request
from flask_restful import Resource, Api, reqparse
from indra import reach
from indra.statements import *
import json
app = Flask(__name__)
api = Api(app)
parser = reqparse.RequestParser()
parser.add_argument('txt')
parser.add_argument('json')
class InputText(Resource):
def post(self):
args = parser.parse_args()
txt = args['txt']
rp = reach.process_text(txt, offline=False)
st = rp.statements
json_statements = {}
json_statements['statements'] = []
for s in st:
s_json = s.to_json()
json_statements['statements'].append(s_json)
json_statements = json.dumps(json_statements)
return json_statements, 201
api.add_resource(InputText, '/parse')
class InputStmtJSON(Resource):
def post(self):
args = parser.parse_args()
print(args)
json_data = args['json']
json_dict = json.loads(json_data)
st = []
for j in json_dict['statements']:
s = Statement.from_json(j)
print(s)
st.append(s)
return 201
api.add_resource(InputStmtJSON, '/load')
if __name__ == '__main__':
app.run(debug=True)
| import json
from bottle import route, run, request, post, default_app
from indra import trips, reach, bel, biopax
from indra.statements import *
@route('/trips/process_text', method='POST')
def trips_process_text():
body = json.load(request.body)
text = body.get('text')
tp = trips.process_text(text)
if tp and tp.statements:
stmts = json.dumps([json.loads(st.to_json()) for st
in tp.statements])
res = {'statements': stmts}
return res
else:
res = {'statements': []}
return res
@route('/reach/process_text', method='POST')
def reach_process_text():
body = json.load(request.body)
text = body.get('text')
rp = reach.process_text(text)
if rp and rp.statements:
stmts = json.dumps([json.loads(st.to_json()) for st
in rp.statements])
res = {'statements': stmts}
return res
else:
res = {'statements': []}
return res
@route('/reach/process_pmc', method='POST')
def reach_process_pmc():
body = json.load(request.body)
pmcid = body.get('pmcid')
rp = reach.process_pmc(pmcid)
if rp and rp.statements:
stmts = json.dumps([json.loads(st.to_json()) for st
in rp.statements])
res = {'statements': stmts}
return res
else:
res = {'statements': []}
return res
if __name__ == '__main__':
app = default_app()
run(app)
| Reimplement using bottle and add 3 endpoints | Reimplement using bottle and add 3 endpoints
| Python | bsd-2-clause | sorgerlab/indra,sorgerlab/indra,sorgerlab/belpy,pvtodorov/indra,bgyori/indra,johnbachman/indra,johnbachman/indra,pvtodorov/indra,johnbachman/belpy,johnbachman/belpy,pvtodorov/indra,bgyori/indra,johnbachman/belpy,bgyori/indra,sorgerlab/indra,pvtodorov/indra,sorgerlab/belpy,sorgerlab/belpy,johnbachman/indra | - from flask import Flask, request
- from flask_restful import Resource, Api, reqparse
- from indra import reach
+ import json
+ from bottle import route, run, request, post, default_app
+ from indra import trips, reach, bel, biopax
from indra.statements import *
- import json
- app = Flask(__name__)
- api = Api(app)
- parser = reqparse.RequestParser()
- parser.add_argument('txt')
- parser.add_argument('json')
- class InputText(Resource):
- def post(self):
- args = parser.parse_args()
- txt = args['txt']
- rp = reach.process_text(txt, offline=False)
- st = rp.statements
- json_statements = {}
- json_statements['statements'] = []
- for s in st:
- s_json = s.to_json()
- json_statements['statements'].append(s_json)
- json_statements = json.dumps(json_statements)
- return json_statements, 201
+ @route('/trips/process_text', method='POST')
+ def trips_process_text():
+ body = json.load(request.body)
+ text = body.get('text')
+ tp = trips.process_text(text)
+ if tp and tp.statements:
+ stmts = json.dumps([json.loads(st.to_json()) for st
+ in tp.statements])
+ res = {'statements': stmts}
+ return res
+ else:
+ res = {'statements': []}
+ return res
- api.add_resource(InputText, '/parse')
+ @route('/reach/process_text', method='POST')
+ def reach_process_text():
+ body = json.load(request.body)
+ text = body.get('text')
+ rp = reach.process_text(text)
+ if rp and rp.statements:
+ stmts = json.dumps([json.loads(st.to_json()) for st
+ in rp.statements])
+ res = {'statements': stmts}
- class InputStmtJSON(Resource):
- def post(self):
- args = parser.parse_args()
- print(args)
- json_data = args['json']
- json_dict = json.loads(json_data)
- st = []
- for j in json_dict['statements']:
- s = Statement.from_json(j)
- print(s)
- st.append(s)
- return 201
+ return res
+ else:
+ res = {'statements': []}
+ return res
- api.add_resource(InputStmtJSON, '/load')
+
+ @route('/reach/process_pmc', method='POST')
+ def reach_process_pmc():
+ body = json.load(request.body)
+ pmcid = body.get('pmcid')
+ rp = reach.process_pmc(pmcid)
+ if rp and rp.statements:
+ stmts = json.dumps([json.loads(st.to_json()) for st
+ in rp.statements])
+ res = {'statements': stmts}
+ return res
+ else:
+ res = {'statements': []}
+ return res
+
if __name__ == '__main__':
- app.run(debug=True)
+ app = default_app()
+ run(app)
| Reimplement using bottle and add 3 endpoints | ## Code Before:
from flask import Flask, request
from flask_restful import Resource, Api, reqparse
from indra import reach
from indra.statements import *
import json
app = Flask(__name__)
api = Api(app)
parser = reqparse.RequestParser()
parser.add_argument('txt')
parser.add_argument('json')
class InputText(Resource):
def post(self):
args = parser.parse_args()
txt = args['txt']
rp = reach.process_text(txt, offline=False)
st = rp.statements
json_statements = {}
json_statements['statements'] = []
for s in st:
s_json = s.to_json()
json_statements['statements'].append(s_json)
json_statements = json.dumps(json_statements)
return json_statements, 201
api.add_resource(InputText, '/parse')
class InputStmtJSON(Resource):
def post(self):
args = parser.parse_args()
print(args)
json_data = args['json']
json_dict = json.loads(json_data)
st = []
for j in json_dict['statements']:
s = Statement.from_json(j)
print(s)
st.append(s)
return 201
api.add_resource(InputStmtJSON, '/load')
if __name__ == '__main__':
app.run(debug=True)
## Instruction:
Reimplement using bottle and add 3 endpoints
## Code After:
import json
from bottle import route, run, request, post, default_app
from indra import trips, reach, bel, biopax
from indra.statements import *
@route('/trips/process_text', method='POST')
def trips_process_text():
body = json.load(request.body)
text = body.get('text')
tp = trips.process_text(text)
if tp and tp.statements:
stmts = json.dumps([json.loads(st.to_json()) for st
in tp.statements])
res = {'statements': stmts}
return res
else:
res = {'statements': []}
return res
@route('/reach/process_text', method='POST')
def reach_process_text():
body = json.load(request.body)
text = body.get('text')
rp = reach.process_text(text)
if rp and rp.statements:
stmts = json.dumps([json.loads(st.to_json()) for st
in rp.statements])
res = {'statements': stmts}
return res
else:
res = {'statements': []}
return res
@route('/reach/process_pmc', method='POST')
def reach_process_pmc():
body = json.load(request.body)
pmcid = body.get('pmcid')
rp = reach.process_pmc(pmcid)
if rp and rp.statements:
stmts = json.dumps([json.loads(st.to_json()) for st
in rp.statements])
res = {'statements': stmts}
return res
else:
res = {'statements': []}
return res
if __name__ == '__main__':
app = default_app()
run(app)
|
e78910c8b9ecf48f96a693dae3c15afa32a12da1 | casexml/apps/phone/views.py | casexml/apps/phone/views.py | from django_digest.decorators import *
from casexml.apps.phone import xml
from casexml.apps.case.models import CommCareCase
from casexml.apps.phone.restore import generate_restore_response
from casexml.apps.phone.models import User
from casexml.apps.case import const
@httpdigest
def restore(request):
user = User.from_django_user(request.user)
restore_id = request.GET.get('since')
return generate_restore_response(user, restore_id)
def xml_for_case(request, case_id, version="1.0"):
"""
Test view to get the xml for a particular case
"""
from django.http import HttpResponse
case = CommCareCase.get(case_id)
return HttpResponse(xml.get_case_xml(case, [const.CASE_ACTION_CREATE,
const.CASE_ACTION_UPDATE],
version), mimetype="text/xml")
| from django.http import HttpResponse
from django_digest.decorators import *
from casexml.apps.phone import xml
from casexml.apps.case.models import CommCareCase
from casexml.apps.phone.restore import generate_restore_response
from casexml.apps.phone.models import User
from casexml.apps.case import const
@httpdigest
def restore(request):
user = User.from_django_user(request.user)
restore_id = request.GET.get('since')
return generate_restore_response(user, restore_id)
def xml_for_case(request, case_id, version="1.0"):
"""
Test view to get the xml for a particular case
"""
case = CommCareCase.get(case_id)
return HttpResponse(xml.get_case_xml(case, [const.CASE_ACTION_CREATE,
const.CASE_ACTION_UPDATE],
version), mimetype="text/xml")
| Revert "moving httpresponse to view" | Revert "moving httpresponse to view"
This reverts commit a6f501bb9de6382e35372996851916adac067fa0.
| Python | bsd-3-clause | SEL-Columbia/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,SEL-Columbia/commcare-hq,qedsoftware/commcare-hq,puttarajubr/commcare-hq,puttarajubr/commcare-hq,puttarajubr/commcare-hq,dimagi/commcare-hq,puttarajubr/commcare-hq,SEL-Columbia/commcare-hq | + from django.http import HttpResponse
from django_digest.decorators import *
from casexml.apps.phone import xml
from casexml.apps.case.models import CommCareCase
from casexml.apps.phone.restore import generate_restore_response
from casexml.apps.phone.models import User
from casexml.apps.case import const
@httpdigest
def restore(request):
user = User.from_django_user(request.user)
restore_id = request.GET.get('since')
return generate_restore_response(user, restore_id)
def xml_for_case(request, case_id, version="1.0"):
"""
Test view to get the xml for a particular case
"""
- from django.http import HttpResponse
case = CommCareCase.get(case_id)
return HttpResponse(xml.get_case_xml(case, [const.CASE_ACTION_CREATE,
const.CASE_ACTION_UPDATE],
version), mimetype="text/xml")
| Revert "moving httpresponse to view" | ## Code Before:
from django_digest.decorators import *
from casexml.apps.phone import xml
from casexml.apps.case.models import CommCareCase
from casexml.apps.phone.restore import generate_restore_response
from casexml.apps.phone.models import User
from casexml.apps.case import const
@httpdigest
def restore(request):
user = User.from_django_user(request.user)
restore_id = request.GET.get('since')
return generate_restore_response(user, restore_id)
def xml_for_case(request, case_id, version="1.0"):
"""
Test view to get the xml for a particular case
"""
from django.http import HttpResponse
case = CommCareCase.get(case_id)
return HttpResponse(xml.get_case_xml(case, [const.CASE_ACTION_CREATE,
const.CASE_ACTION_UPDATE],
version), mimetype="text/xml")
## Instruction:
Revert "moving httpresponse to view"
## Code After:
from django.http import HttpResponse
from django_digest.decorators import *
from casexml.apps.phone import xml
from casexml.apps.case.models import CommCareCase
from casexml.apps.phone.restore import generate_restore_response
from casexml.apps.phone.models import User
from casexml.apps.case import const
@httpdigest
def restore(request):
user = User.from_django_user(request.user)
restore_id = request.GET.get('since')
return generate_restore_response(user, restore_id)
def xml_for_case(request, case_id, version="1.0"):
"""
Test view to get the xml for a particular case
"""
case = CommCareCase.get(case_id)
return HttpResponse(xml.get_case_xml(case, [const.CASE_ACTION_CREATE,
const.CASE_ACTION_UPDATE],
version), mimetype="text/xml")
|
63f6e4d50116d5ca2bfc82c1c608e08040055b5e | subdue/core/__init__.py | subdue/core/__init__.py | __all__ = [
'color',
'BANNER',
'DEFAULT_DRIVER_CODE'
'die',
'verbose',
'use_colors',
'set_color_policy',
]
import sys as _sys
from . import color as _color
BANNER = """\
_ _
___ _ _| |__ __| |_ _ ___
/ __| | | | '_ \ / _` | | | |/ _ \\
\__ \ |_| | |_) | (_| | |_| | __/
|___/\__,_|_.__/ \__,_|\__,_|\___|
"""
DEFAULT_DRIVER_CODE = """\
#!/usr/bin/env python
from subdue.sub import main
main()
"""
verbose = False
def set_color_policy(policy):
_color.color_policy = policy
def die(msg):
_sys.stderr.write(msg)
_sys.stderr.write("\n")
_sys.stderr.flush()
_sys.exit(1)
| __all__ = [
'BANNER',
'DEFAULT_DRIVER_CODE'
'die',
'verbose',
'set_color_policy',
]
import sys as _sys
from . import color as _color
BANNER = """\
_ _
___ _ _| |__ __| |_ _ ___
/ __| | | | '_ \ / _` | | | |/ _ \\
\__ \ |_| | |_) | (_| | |_| | __/
|___/\__,_|_.__/ \__,_|\__,_|\___|
"""
DEFAULT_DRIVER_CODE = """\
#!/usr/bin/env python
from subdue.sub import main
main()
"""
verbose = False
def set_color_policy(policy):
_color.color_policy = policy
def die(msg):
_sys.stderr.write(msg)
_sys.stderr.write("\n")
_sys.stderr.flush()
_sys.exit(1)
| Remove old exports from subdue.core | Remove old exports from subdue.core
| Python | mit | jdevera/subdue | __all__ = [
- 'color',
'BANNER',
'DEFAULT_DRIVER_CODE'
'die',
'verbose',
- 'use_colors',
'set_color_policy',
]
import sys as _sys
from . import color as _color
BANNER = """\
_ _
___ _ _| |__ __| |_ _ ___
/ __| | | | '_ \ / _` | | | |/ _ \\
\__ \ |_| | |_) | (_| | |_| | __/
|___/\__,_|_.__/ \__,_|\__,_|\___|
"""
DEFAULT_DRIVER_CODE = """\
#!/usr/bin/env python
from subdue.sub import main
main()
"""
verbose = False
def set_color_policy(policy):
_color.color_policy = policy
def die(msg):
_sys.stderr.write(msg)
_sys.stderr.write("\n")
_sys.stderr.flush()
_sys.exit(1)
| Remove old exports from subdue.core | ## Code Before:
__all__ = [
'color',
'BANNER',
'DEFAULT_DRIVER_CODE'
'die',
'verbose',
'use_colors',
'set_color_policy',
]
import sys as _sys
from . import color as _color
BANNER = """\
_ _
___ _ _| |__ __| |_ _ ___
/ __| | | | '_ \ / _` | | | |/ _ \\
\__ \ |_| | |_) | (_| | |_| | __/
|___/\__,_|_.__/ \__,_|\__,_|\___|
"""
DEFAULT_DRIVER_CODE = """\
#!/usr/bin/env python
from subdue.sub import main
main()
"""
verbose = False
def set_color_policy(policy):
_color.color_policy = policy
def die(msg):
_sys.stderr.write(msg)
_sys.stderr.write("\n")
_sys.stderr.flush()
_sys.exit(1)
## Instruction:
Remove old exports from subdue.core
## Code After:
__all__ = [
'BANNER',
'DEFAULT_DRIVER_CODE'
'die',
'verbose',
'set_color_policy',
]
import sys as _sys
from . import color as _color
BANNER = """\
_ _
___ _ _| |__ __| |_ _ ___
/ __| | | | '_ \ / _` | | | |/ _ \\
\__ \ |_| | |_) | (_| | |_| | __/
|___/\__,_|_.__/ \__,_|\__,_|\___|
"""
DEFAULT_DRIVER_CODE = """\
#!/usr/bin/env python
from subdue.sub import main
main()
"""
verbose = False
def set_color_policy(policy):
_color.color_policy = policy
def die(msg):
_sys.stderr.write(msg)
_sys.stderr.write("\n")
_sys.stderr.flush()
_sys.exit(1)
|
459bf08b9fe4ae5a879a138bd2497abb23bf5910 | modules/expansion/cve.py | modules/expansion/cve.py | import json
import requests
misperrors = {'error': 'Error'}
mispattributes = {'input': ['vulnerability'], 'output': ['']}
moduleinfo = {'version': '0.1', 'author': 'Alexandre Dulaunoy', 'description': 'An expansion hover module to expand information about CVE id.', 'module-type': ['hover']}
moduleconfig = []
cveapi_url = 'https://cve.circl.lu/api/cve/'
def handler(q=False):
if q is False:
return False
print (q)
request = json.loads(q)
if not request.get('vulnerability'):
misperrors['error'] = 'Vulnerability id missing'
return misperrors
r = requests.get(cveapi_url+request.get('vulnerability'))
if r.status_code == 200:
vulnerability = json.loads(r.text)
else:
misperrors['error'] = 'cve.circl.lu API not accessible'
return misperrors['error']
return vulnerability
def introspection():
return mispattributes
def version():
moduleinfo['config'] = moduleconfig
return moduleinfo
| import json
import requests
misperrors = {'error': 'Error'}
mispattributes = {'input': ['vulnerability'], 'output': ['text']}
moduleinfo = {'version': '0.2', 'author': 'Alexandre Dulaunoy', 'description': 'An expansion hover module to expand information about CVE id.', 'module-type': ['hover']}
moduleconfig = []
cveapi_url = 'https://cve.circl.lu/api/cve/'
def handler(q=False):
if q is False:
return False
print (q)
request = json.loads(q)
if not request.get('vulnerability'):
misperrors['error'] = 'Vulnerability id missing'
return misperrors
r = requests.get(cveapi_url+request.get('vulnerability'))
if r.status_code == 200:
vulnerability = json.loads(r.text)
if vulnerability.get('summary'):
summary = vulnerability['summary']
else:
misperrors['error'] = 'cve.circl.lu API not accessible'
return misperrors['error']
r = {'results': [{'types': mispattributes['output'], 'values': summary}]}
return r
def introspection():
return mispattributes
def version():
moduleinfo['config'] = moduleconfig
return moduleinfo
| Return a text attribute for an hover only module | Return a text attribute for an hover only module
| Python | agpl-3.0 | VirusTotal/misp-modules,MISP/misp-modules,MISP/misp-modules,amuehlem/misp-modules,MISP/misp-modules,Rafiot/misp-modules,Rafiot/misp-modules,amuehlem/misp-modules,Rafiot/misp-modules,amuehlem/misp-modules,VirusTotal/misp-modules,VirusTotal/misp-modules | import json
import requests
misperrors = {'error': 'Error'}
- mispattributes = {'input': ['vulnerability'], 'output': ['']}
+ mispattributes = {'input': ['vulnerability'], 'output': ['text']}
- moduleinfo = {'version': '0.1', 'author': 'Alexandre Dulaunoy', 'description': 'An expansion hover module to expand information about CVE id.', 'module-type': ['hover']}
+ moduleinfo = {'version': '0.2', 'author': 'Alexandre Dulaunoy', 'description': 'An expansion hover module to expand information about CVE id.', 'module-type': ['hover']}
moduleconfig = []
cveapi_url = 'https://cve.circl.lu/api/cve/'
def handler(q=False):
if q is False:
return False
print (q)
request = json.loads(q)
if not request.get('vulnerability'):
misperrors['error'] = 'Vulnerability id missing'
return misperrors
r = requests.get(cveapi_url+request.get('vulnerability'))
if r.status_code == 200:
vulnerability = json.loads(r.text)
+ if vulnerability.get('summary'):
+ summary = vulnerability['summary']
else:
misperrors['error'] = 'cve.circl.lu API not accessible'
return misperrors['error']
- return vulnerability
+ r = {'results': [{'types': mispattributes['output'], 'values': summary}]}
+ return r
def introspection():
return mispattributes
def version():
moduleinfo['config'] = moduleconfig
return moduleinfo
| Return a text attribute for an hover only module | ## Code Before:
import json
import requests
misperrors = {'error': 'Error'}
mispattributes = {'input': ['vulnerability'], 'output': ['']}
moduleinfo = {'version': '0.1', 'author': 'Alexandre Dulaunoy', 'description': 'An expansion hover module to expand information about CVE id.', 'module-type': ['hover']}
moduleconfig = []
cveapi_url = 'https://cve.circl.lu/api/cve/'
def handler(q=False):
if q is False:
return False
print (q)
request = json.loads(q)
if not request.get('vulnerability'):
misperrors['error'] = 'Vulnerability id missing'
return misperrors
r = requests.get(cveapi_url+request.get('vulnerability'))
if r.status_code == 200:
vulnerability = json.loads(r.text)
else:
misperrors['error'] = 'cve.circl.lu API not accessible'
return misperrors['error']
return vulnerability
def introspection():
return mispattributes
def version():
moduleinfo['config'] = moduleconfig
return moduleinfo
## Instruction:
Return a text attribute for an hover only module
## Code After:
import json
import requests
misperrors = {'error': 'Error'}
mispattributes = {'input': ['vulnerability'], 'output': ['text']}
moduleinfo = {'version': '0.2', 'author': 'Alexandre Dulaunoy', 'description': 'An expansion hover module to expand information about CVE id.', 'module-type': ['hover']}
moduleconfig = []
cveapi_url = 'https://cve.circl.lu/api/cve/'
def handler(q=False):
if q is False:
return False
print (q)
request = json.loads(q)
if not request.get('vulnerability'):
misperrors['error'] = 'Vulnerability id missing'
return misperrors
r = requests.get(cveapi_url+request.get('vulnerability'))
if r.status_code == 200:
vulnerability = json.loads(r.text)
if vulnerability.get('summary'):
summary = vulnerability['summary']
else:
misperrors['error'] = 'cve.circl.lu API not accessible'
return misperrors['error']
r = {'results': [{'types': mispattributes['output'], 'values': summary}]}
return r
def introspection():
return mispattributes
def version():
moduleinfo['config'] = moduleconfig
return moduleinfo
|
03380a1042443465d6f1d74afb5fd120dbc3379b | manage.py | manage.py | from manager import Manager
manager = Manager()
@manager.command
def build(threads=1):
print("Starting a build with %d threads ..." % threads)
@manager.command
def clean():
pass
if __name__ == '__main__':
manager.main()
| from manager import Manager
from multiprocessing import Pool
manager = Manager()
def func(period):
from time import sleep
sleep(period)
@manager.command
def build(threads=1):
pool = Pool(threads)
print("Starting a build with %d threads ..." % threads)
pool.map(func, [1, 1, 1, 1, 1])
@manager.command
def clean():
pass
if __name__ == '__main__':
manager.main()
| Add parallelizing code to build | Add parallelizing code to build
| Python | mit | tanayseven/personal_website,tanayseven/personal_website,tanayseven/personal_website,tanayseven/personal_website | from manager import Manager
+ from multiprocessing import Pool
manager = Manager()
+ def func(period):
+ from time import sleep
+ sleep(period)
+
+
@manager.command
def build(threads=1):
+ pool = Pool(threads)
print("Starting a build with %d threads ..." % threads)
+ pool.map(func, [1, 1, 1, 1, 1])
+
@manager.command
def clean():
pass
+
if __name__ == '__main__':
manager.main()
| Add parallelizing code to build | ## Code Before:
from manager import Manager
manager = Manager()
@manager.command
def build(threads=1):
print("Starting a build with %d threads ..." % threads)
@manager.command
def clean():
pass
if __name__ == '__main__':
manager.main()
## Instruction:
Add parallelizing code to build
## Code After:
from manager import Manager
from multiprocessing import Pool
manager = Manager()
def func(period):
from time import sleep
sleep(period)
@manager.command
def build(threads=1):
pool = Pool(threads)
print("Starting a build with %d threads ..." % threads)
pool.map(func, [1, 1, 1, 1, 1])
@manager.command
def clean():
pass
if __name__ == '__main__':
manager.main()
|
a5003b6f45d262923a1c00bd9a9c1addb3854178 | lapostesdk/apis/apibase.py | lapostesdk/apis/apibase.py | import requests
from importlib import import_module
class ApiBase(object):
def __init__(self, api_key, product, version='v1', entity=None):
self.product = product
self.version = version
self.entity = entity
self.api_url = 'https://api.laposte.fr/%(product)s/%(version)s/' % {
'product': self.product,
'version': self.version}
self.headers = {'X-Okapi-Key': api_key}
def get(self, resource, params={}):
response = self._get(resource, params)
if self.entity is None:
return response
module = import_module('lapostesdk.entities')
obj = getattr(module, self.entity)
instance = obj()
instance.hydrate(response)
return instance
def _get(self, resource, params={}):
r = requests.get(self.api_url + resource, params=params, headers=self.headers)
return r.json()
| import requests
from importlib import import_module
class ApiBase(object):
def __init__(self, api_key, product, version='v1', entity=None):
self.product = product
self.version = version
self.entity = entity
self.api_url = 'https://api.laposte.fr/%(product)s/%(version)s/' % {
'product': self.product,
'version': self.version}
self.headers = {'X-Okapi-Key': api_key}
def get(self, resource, params={}):
response = self._get(resource, params)
if self.entity is None:
return response
return self.create_object(response, self.entity)
def _get(self, resource, params={}):
r = requests.get(self.api_url + resource, params=params, headers=self.headers)
return r.json()
def create_object(self, response, entity):
module = import_module('lapostesdk.entities')
obj = getattr(module, self.entity)
instance = obj()
instance.hydrate(response)
return instance
| Move object creation outside of get method | Move object creation outside of get method
| Python | mit | geelweb/laposte-python-sdk | import requests
from importlib import import_module
class ApiBase(object):
def __init__(self, api_key, product, version='v1', entity=None):
self.product = product
self.version = version
self.entity = entity
self.api_url = 'https://api.laposte.fr/%(product)s/%(version)s/' % {
'product': self.product,
'version': self.version}
self.headers = {'X-Okapi-Key': api_key}
def get(self, resource, params={}):
response = self._get(resource, params)
if self.entity is None:
return response
+ return self.create_object(response, self.entity)
+
+ def _get(self, resource, params={}):
+ r = requests.get(self.api_url + resource, params=params, headers=self.headers)
+ return r.json()
+
+ def create_object(self, response, entity):
module = import_module('lapostesdk.entities')
obj = getattr(module, self.entity)
instance = obj()
instance.hydrate(response)
return instance
- def _get(self, resource, params={}):
- r = requests.get(self.api_url + resource, params=params, headers=self.headers)
- return r.json()
| Move object creation outside of get method | ## Code Before:
import requests
from importlib import import_module
class ApiBase(object):
def __init__(self, api_key, product, version='v1', entity=None):
self.product = product
self.version = version
self.entity = entity
self.api_url = 'https://api.laposte.fr/%(product)s/%(version)s/' % {
'product': self.product,
'version': self.version}
self.headers = {'X-Okapi-Key': api_key}
def get(self, resource, params={}):
response = self._get(resource, params)
if self.entity is None:
return response
module = import_module('lapostesdk.entities')
obj = getattr(module, self.entity)
instance = obj()
instance.hydrate(response)
return instance
def _get(self, resource, params={}):
r = requests.get(self.api_url + resource, params=params, headers=self.headers)
return r.json()
## Instruction:
Move object creation outside of get method
## Code After:
import requests
from importlib import import_module
class ApiBase(object):
def __init__(self, api_key, product, version='v1', entity=None):
self.product = product
self.version = version
self.entity = entity
self.api_url = 'https://api.laposte.fr/%(product)s/%(version)s/' % {
'product': self.product,
'version': self.version}
self.headers = {'X-Okapi-Key': api_key}
def get(self, resource, params={}):
response = self._get(resource, params)
if self.entity is None:
return response
return self.create_object(response, self.entity)
def _get(self, resource, params={}):
r = requests.get(self.api_url + resource, params=params, headers=self.headers)
return r.json()
def create_object(self, response, entity):
module = import_module('lapostesdk.entities')
obj = getattr(module, self.entity)
instance = obj()
instance.hydrate(response)
return instance
|
1b385ce127f0a1802b0effa0054b44f58b3317b0 | {{cookiecutter.repo_name}}/{{cookiecutter.repo_name}}/accounts/urls.py | {{cookiecutter.repo_name}}/{{cookiecutter.repo_name}}/accounts/urls.py | from django.contrib.auth import views
from django.urls import path, re_path
from accounts.forms import LoginForm, PasswordResetForm, SetPasswordForm
urlpatterns = [
path(
"login/",
views.LoginView.as_view(
template_name="accounts/login.html", authentication_form=LoginForm
),
name="login",
),
path("logout/", views.LogoutView.as_view(), name="logout"),
# Password reset
path(
"account/password_reset/",
views.PasswordResetView.as_view(form_class=PasswordResetForm),
name="password_reset",
),
path(
"account/password_reset/done/",
views.PasswordResetDoneView.as_view(),
name="password_reset_done",
),
re_path(
r"^account/reset/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$",
views.PasswordResetConfirmView.as_view(form_class=SetPasswordForm),
name="password_reset_confirm",
),
path(
"account/reset/done/",
views.PasswordResetCompleteView.as_view(),
name="password_reset_complete",
),
]
| from django.contrib.auth import views
from django.urls import path
from accounts.forms import LoginForm, PasswordResetForm, SetPasswordForm
urlpatterns = [
path(
"login/",
views.LoginView.as_view(
template_name="accounts/login.html", authentication_form=LoginForm
),
name="login",
),
path("logout/", views.LogoutView.as_view(), name="logout"),
# Password reset
path(
"account/password_reset/",
views.PasswordResetView.as_view(form_class=PasswordResetForm),
name="password_reset",
),
path(
"account/password_reset/done/",
views.PasswordResetDoneView.as_view(),
name="password_reset_done",
),
path(
r"account/reset/<uidb64>/<token>/",
views.PasswordResetConfirmView.as_view(form_class=SetPasswordForm),
name="password_reset_confirm",
),
path(
"account/reset/done/",
views.PasswordResetCompleteView.as_view(),
name="password_reset_complete",
),
]
| Fix webapp password reset link | DEVOPS-42: Fix webapp password reset link
| Python | isc | thorgate/django-project-template,thorgate/django-project-template,thorgate/django-project-template,thorgate/django-project-template,thorgate/django-project-template | from django.contrib.auth import views
- from django.urls import path, re_path
+ from django.urls import path
from accounts.forms import LoginForm, PasswordResetForm, SetPasswordForm
urlpatterns = [
path(
"login/",
views.LoginView.as_view(
template_name="accounts/login.html", authentication_form=LoginForm
),
name="login",
),
path("logout/", views.LogoutView.as_view(), name="logout"),
# Password reset
path(
"account/password_reset/",
views.PasswordResetView.as_view(form_class=PasswordResetForm),
name="password_reset",
),
path(
"account/password_reset/done/",
views.PasswordResetDoneView.as_view(),
name="password_reset_done",
),
- re_path(
+ path(
- r"^account/reset/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$",
+ r"account/reset/<uidb64>/<token>/",
views.PasswordResetConfirmView.as_view(form_class=SetPasswordForm),
name="password_reset_confirm",
),
path(
"account/reset/done/",
views.PasswordResetCompleteView.as_view(),
name="password_reset_complete",
),
]
| Fix webapp password reset link | ## Code Before:
from django.contrib.auth import views
from django.urls import path, re_path
from accounts.forms import LoginForm, PasswordResetForm, SetPasswordForm
urlpatterns = [
path(
"login/",
views.LoginView.as_view(
template_name="accounts/login.html", authentication_form=LoginForm
),
name="login",
),
path("logout/", views.LogoutView.as_view(), name="logout"),
# Password reset
path(
"account/password_reset/",
views.PasswordResetView.as_view(form_class=PasswordResetForm),
name="password_reset",
),
path(
"account/password_reset/done/",
views.PasswordResetDoneView.as_view(),
name="password_reset_done",
),
re_path(
r"^account/reset/(?P<uidb64>[0-9A-Za-z_\-]+)/(?P<token>[0-9A-Za-z]{1,13}-[0-9A-Za-z]{1,20})/$",
views.PasswordResetConfirmView.as_view(form_class=SetPasswordForm),
name="password_reset_confirm",
),
path(
"account/reset/done/",
views.PasswordResetCompleteView.as_view(),
name="password_reset_complete",
),
]
## Instruction:
Fix webapp password reset link
## Code After:
from django.contrib.auth import views
from django.urls import path
from accounts.forms import LoginForm, PasswordResetForm, SetPasswordForm
urlpatterns = [
path(
"login/",
views.LoginView.as_view(
template_name="accounts/login.html", authentication_form=LoginForm
),
name="login",
),
path("logout/", views.LogoutView.as_view(), name="logout"),
# Password reset
path(
"account/password_reset/",
views.PasswordResetView.as_view(form_class=PasswordResetForm),
name="password_reset",
),
path(
"account/password_reset/done/",
views.PasswordResetDoneView.as_view(),
name="password_reset_done",
),
path(
r"account/reset/<uidb64>/<token>/",
views.PasswordResetConfirmView.as_view(form_class=SetPasswordForm),
name="password_reset_confirm",
),
path(
"account/reset/done/",
views.PasswordResetCompleteView.as_view(),
name="password_reset_complete",
),
]
|
9b10f600b5611380f72fe2aeacfe2ee6f02e4e3a | kicad_footprint_load.py | kicad_footprint_load.py | import pcbnew
import sys
import os
pretties = []
for dirname, dirnames, filenames in os.walk(sys.argv[1]):
# don't go into any .git directories.
if '.git' in dirnames:
dirnames.remove('.git')
for filename in filenames:
if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'):
pretties.append(os.path.realpath(dirname))
break
src_plugin = pcbnew.IO_MGR.PluginFind(1)
for libpath in pretties:
#Ignore paths with unicode as KiCad can't deal with them in enumerate
list_of_footprints = src_plugin.FootprintEnumerate(libpath, False)
| import pcbnew
import sys
import os
pretties = []
for dirname, dirnames, filenames in os.walk(sys.argv[1]):
# don't go into any .git directories.
if '.git' in dirnames:
dirnames.remove('.git')
for filename in filenames:
if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'):
pretties.append(os.path.realpath(dirname))
break
src_plugin = pcbnew.IO_MGR.PluginFind(1)
for libpath in pretties:
list_of_footprints = src_plugin.FootprintEnumerate(libpath)
| Switch to old invocation of FootprintEnumerate | Switch to old invocation of FootprintEnumerate
| Python | mit | monostable/haskell-kicad-data,monostable/haskell-kicad-data,kasbah/haskell-kicad-data | import pcbnew
import sys
import os
pretties = []
for dirname, dirnames, filenames in os.walk(sys.argv[1]):
# don't go into any .git directories.
if '.git' in dirnames:
dirnames.remove('.git')
for filename in filenames:
if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'):
pretties.append(os.path.realpath(dirname))
break
src_plugin = pcbnew.IO_MGR.PluginFind(1)
for libpath in pretties:
- #Ignore paths with unicode as KiCad can't deal with them in enumerate
- list_of_footprints = src_plugin.FootprintEnumerate(libpath, False)
+ list_of_footprints = src_plugin.FootprintEnumerate(libpath)
| Switch to old invocation of FootprintEnumerate | ## Code Before:
import pcbnew
import sys
import os
pretties = []
for dirname, dirnames, filenames in os.walk(sys.argv[1]):
# don't go into any .git directories.
if '.git' in dirnames:
dirnames.remove('.git')
for filename in filenames:
if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'):
pretties.append(os.path.realpath(dirname))
break
src_plugin = pcbnew.IO_MGR.PluginFind(1)
for libpath in pretties:
#Ignore paths with unicode as KiCad can't deal with them in enumerate
list_of_footprints = src_plugin.FootprintEnumerate(libpath, False)
## Instruction:
Switch to old invocation of FootprintEnumerate
## Code After:
import pcbnew
import sys
import os
pretties = []
for dirname, dirnames, filenames in os.walk(sys.argv[1]):
# don't go into any .git directories.
if '.git' in dirnames:
dirnames.remove('.git')
for filename in filenames:
if (not os.path.isdir(filename)) and (os.path.splitext(filename)[-1] == '.kicad_mod'):
pretties.append(os.path.realpath(dirname))
break
src_plugin = pcbnew.IO_MGR.PluginFind(1)
for libpath in pretties:
list_of_footprints = src_plugin.FootprintEnumerate(libpath)
|
f5fd283497afb5030632108ce692e8acde526188 | datalake_ingester/reporter.py | datalake_ingester/reporter.py | import boto.sns
import simplejson as json
import logging
from memoized_property import memoized_property
import os
from datalake_common.errors import InsufficientConfiguration
class SNSReporter(object):
'''report ingestion events to SNS'''
def __init__(self, report_key):
self.report_key = report_key
self.logger = logging.getLogger(self._log_name)
@classmethod
def from_config(cls):
report_key = os.environ.get('DATALAKE_REPORT_KEY')
if report_key is None:
raise InsufficientConfiguration('Please configure a report_key')
return cls(report_key)
@property
def _log_name(self):
return self.report_key.split(':')[-1]
@memoized_property
def _connection(self):
region = os.environ.get('AWS_REGION')
if region:
return boto.sns.connect_to_region(region)
else:
return boto.connect_sns()
def report(self, ingestion_report):
message = json.dumps(ingestion_report)
self.logger.info('REPORTING: %s', message)
self._connection.publish(topic=self.report_key, message=message)
| import boto.sns
import simplejson as json
import logging
from memoized_property import memoized_property
import os
class SNSReporter(object):
'''report ingestion events to SNS'''
def __init__(self, report_key):
self.report_key = report_key
self.logger = logging.getLogger(self._log_name)
@classmethod
def from_config(cls):
report_key = os.environ.get('DATALAKE_REPORT_KEY')
if report_key is None:
return None
return cls(report_key)
@property
def _log_name(self):
return self.report_key.split(':')[-1]
@memoized_property
def _connection(self):
region = os.environ.get('AWS_REGION')
if region:
return boto.sns.connect_to_region(region)
else:
return boto.connect_sns()
def report(self, ingestion_report):
message = json.dumps(ingestion_report)
self.logger.info('REPORTING: %s', message)
self._connection.publish(topic=self.report_key, message=message)
| Allow the ingester to work without a report key | Allow the ingester to work without a report key
| Python | apache-2.0 | planetlabs/datalake-ingester,planetlabs/atl,planetlabs/datalake,planetlabs/datalake,planetlabs/datalake,planetlabs/datalake | import boto.sns
import simplejson as json
import logging
from memoized_property import memoized_property
import os
- from datalake_common.errors import InsufficientConfiguration
-
class SNSReporter(object):
'''report ingestion events to SNS'''
def __init__(self, report_key):
self.report_key = report_key
self.logger = logging.getLogger(self._log_name)
@classmethod
def from_config(cls):
report_key = os.environ.get('DATALAKE_REPORT_KEY')
if report_key is None:
- raise InsufficientConfiguration('Please configure a report_key')
+ return None
return cls(report_key)
@property
def _log_name(self):
return self.report_key.split(':')[-1]
@memoized_property
def _connection(self):
region = os.environ.get('AWS_REGION')
if region:
return boto.sns.connect_to_region(region)
else:
return boto.connect_sns()
def report(self, ingestion_report):
message = json.dumps(ingestion_report)
self.logger.info('REPORTING: %s', message)
self._connection.publish(topic=self.report_key, message=message)
| Allow the ingester to work without a report key | ## Code Before:
import boto.sns
import simplejson as json
import logging
from memoized_property import memoized_property
import os
from datalake_common.errors import InsufficientConfiguration
class SNSReporter(object):
'''report ingestion events to SNS'''
def __init__(self, report_key):
self.report_key = report_key
self.logger = logging.getLogger(self._log_name)
@classmethod
def from_config(cls):
report_key = os.environ.get('DATALAKE_REPORT_KEY')
if report_key is None:
raise InsufficientConfiguration('Please configure a report_key')
return cls(report_key)
@property
def _log_name(self):
return self.report_key.split(':')[-1]
@memoized_property
def _connection(self):
region = os.environ.get('AWS_REGION')
if region:
return boto.sns.connect_to_region(region)
else:
return boto.connect_sns()
def report(self, ingestion_report):
message = json.dumps(ingestion_report)
self.logger.info('REPORTING: %s', message)
self._connection.publish(topic=self.report_key, message=message)
## Instruction:
Allow the ingester to work without a report key
## Code After:
import boto.sns
import simplejson as json
import logging
from memoized_property import memoized_property
import os
class SNSReporter(object):
'''report ingestion events to SNS'''
def __init__(self, report_key):
self.report_key = report_key
self.logger = logging.getLogger(self._log_name)
@classmethod
def from_config(cls):
report_key = os.environ.get('DATALAKE_REPORT_KEY')
if report_key is None:
return None
return cls(report_key)
@property
def _log_name(self):
return self.report_key.split(':')[-1]
@memoized_property
def _connection(self):
region = os.environ.get('AWS_REGION')
if region:
return boto.sns.connect_to_region(region)
else:
return boto.connect_sns()
def report(self, ingestion_report):
message = json.dumps(ingestion_report)
self.logger.info('REPORTING: %s', message)
self._connection.publish(topic=self.report_key, message=message)
|
f45fc8854647754b24df5f9601920368cd2d3c49 | tests/chainerx_tests/unit_tests/test_cuda.py | tests/chainerx_tests/unit_tests/test_cuda.py | import pytest
from chainerx import _cuda
try:
import cupy
except Exception:
cupy = None
class CupyTestMemoryHook(cupy.cuda.memory_hook.MemoryHook):
name = 'CupyTestMemoryHook'
def __init__(self):
self.used_bytes = 0
self.acquired_bytes = 0
def alloc_preprocess(self, **kwargs):
self.acquired_bytes += kwargs['mem_size']
def malloc_preprocess(self, **kwargs):
self.used_bytes += kwargs['mem_size']
@pytest.mark.cuda()
def test_cupy_share_allocator():
with CupyTestMemoryHook() as hook:
cp_allocated = cupy.arange(10)
used_bytes = hook.used_bytes
acquired_bytes = hook.acquired_bytes
# Create a new array after changing the allocator to the memory pool
# of ChainerX and make sure that no additional memory has been
# allocated by CuPy.
_cuda.cupy_share_allocator()
chx_allocated = cupy.arange(10)
cupy.testing.assert_array_equal(cp_allocated, chx_allocated)
assert used_bytes == hook.used_bytes
assert acquired_bytes == hook.acquired_bytes
| import pytest
from chainerx import _cuda
try:
import cupy
except Exception:
cupy = None
class CupyTestMemoryHook(cupy.cuda.memory_hook.MemoryHook):
name = 'CupyTestMemoryHook'
def __init__(self):
self.used_bytes = 0
self.acquired_bytes = 0
def alloc_preprocess(self, **kwargs):
self.acquired_bytes += kwargs['mem_size']
def malloc_preprocess(self, **kwargs):
self.used_bytes += kwargs['mem_size']
@pytest.mark.cuda()
def test_cupy_share_allocator():
with CupyTestMemoryHook() as hook:
cp_allocated = cupy.arange(10)
used_bytes = hook.used_bytes
acquired_bytes = hook.acquired_bytes
assert used_bytes > 0
assert acquired_bytes > 0
# Create a new array after changing the allocator to the memory pool
# of ChainerX and make sure that no additional memory has been
# allocated by CuPy.
_cuda.cupy_share_allocator()
chx_allocated = cupy.arange(10)
cupy.testing.assert_array_equal(cp_allocated, chx_allocated)
assert used_bytes == hook.used_bytes
assert acquired_bytes == hook.acquired_bytes
| Add safety checks in test | Add safety checks in test
| Python | mit | wkentaro/chainer,hvy/chainer,niboshi/chainer,okuta/chainer,chainer/chainer,wkentaro/chainer,chainer/chainer,keisuke-umezawa/chainer,keisuke-umezawa/chainer,hvy/chainer,pfnet/chainer,hvy/chainer,chainer/chainer,keisuke-umezawa/chainer,okuta/chainer,chainer/chainer,tkerola/chainer,keisuke-umezawa/chainer,wkentaro/chainer,niboshi/chainer,hvy/chainer,okuta/chainer,niboshi/chainer,niboshi/chainer,okuta/chainer,wkentaro/chainer | import pytest
from chainerx import _cuda
try:
import cupy
except Exception:
cupy = None
class CupyTestMemoryHook(cupy.cuda.memory_hook.MemoryHook):
name = 'CupyTestMemoryHook'
def __init__(self):
self.used_bytes = 0
self.acquired_bytes = 0
def alloc_preprocess(self, **kwargs):
self.acquired_bytes += kwargs['mem_size']
def malloc_preprocess(self, **kwargs):
self.used_bytes += kwargs['mem_size']
@pytest.mark.cuda()
def test_cupy_share_allocator():
with CupyTestMemoryHook() as hook:
cp_allocated = cupy.arange(10)
used_bytes = hook.used_bytes
acquired_bytes = hook.acquired_bytes
+ assert used_bytes > 0
+ assert acquired_bytes > 0
# Create a new array after changing the allocator to the memory pool
# of ChainerX and make sure that no additional memory has been
# allocated by CuPy.
_cuda.cupy_share_allocator()
chx_allocated = cupy.arange(10)
cupy.testing.assert_array_equal(cp_allocated, chx_allocated)
assert used_bytes == hook.used_bytes
assert acquired_bytes == hook.acquired_bytes
| Add safety checks in test | ## Code Before:
import pytest
from chainerx import _cuda
try:
import cupy
except Exception:
cupy = None
class CupyTestMemoryHook(cupy.cuda.memory_hook.MemoryHook):
name = 'CupyTestMemoryHook'
def __init__(self):
self.used_bytes = 0
self.acquired_bytes = 0
def alloc_preprocess(self, **kwargs):
self.acquired_bytes += kwargs['mem_size']
def malloc_preprocess(self, **kwargs):
self.used_bytes += kwargs['mem_size']
@pytest.mark.cuda()
def test_cupy_share_allocator():
with CupyTestMemoryHook() as hook:
cp_allocated = cupy.arange(10)
used_bytes = hook.used_bytes
acquired_bytes = hook.acquired_bytes
# Create a new array after changing the allocator to the memory pool
# of ChainerX and make sure that no additional memory has been
# allocated by CuPy.
_cuda.cupy_share_allocator()
chx_allocated = cupy.arange(10)
cupy.testing.assert_array_equal(cp_allocated, chx_allocated)
assert used_bytes == hook.used_bytes
assert acquired_bytes == hook.acquired_bytes
## Instruction:
Add safety checks in test
## Code After:
import pytest
from chainerx import _cuda
try:
import cupy
except Exception:
cupy = None
class CupyTestMemoryHook(cupy.cuda.memory_hook.MemoryHook):
name = 'CupyTestMemoryHook'
def __init__(self):
self.used_bytes = 0
self.acquired_bytes = 0
def alloc_preprocess(self, **kwargs):
self.acquired_bytes += kwargs['mem_size']
def malloc_preprocess(self, **kwargs):
self.used_bytes += kwargs['mem_size']
@pytest.mark.cuda()
def test_cupy_share_allocator():
with CupyTestMemoryHook() as hook:
cp_allocated = cupy.arange(10)
used_bytes = hook.used_bytes
acquired_bytes = hook.acquired_bytes
assert used_bytes > 0
assert acquired_bytes > 0
# Create a new array after changing the allocator to the memory pool
# of ChainerX and make sure that no additional memory has been
# allocated by CuPy.
_cuda.cupy_share_allocator()
chx_allocated = cupy.arange(10)
cupy.testing.assert_array_equal(cp_allocated, chx_allocated)
assert used_bytes == hook.used_bytes
assert acquired_bytes == hook.acquired_bytes
|
429f38497da0fd520e5bc5bd82e6d4ed5a405521 | real_estate_agency/real_estate_agency/views.py | real_estate_agency/real_estate_agency/views.py | from django.shortcuts import render, render_to_response
from django.template import RequestContext
from new_buildings.models import Builder, ResidentalComplex, NewApartment
from new_buildings.forms import SearchForm
from feedback.models import Feedback
def corporation_benefit_plan(request):
return render(request, 'corporation_benefit_plan.html')
def index(request):
# Only 2 requests to DB
feedbacks = Feedback.objects.all()[:4].select_related().prefetch_related('social_media_links')
# Only 2 requests to DB
residental_complexes = ResidentalComplex.objects.filter(
is_popular=True).prefetch_related('type_of_complex')
context = {
'feedbacks': feedbacks,
'form': SearchForm,
'residental_complexes': residental_complexes,
}
return render(request,
'index.html',
context,
)
def privacy_policy(request):
return render(request, 'privacy_policy.html')
def thanks(request):
return render(request, 'thanks.html')
| from django.shortcuts import render
from new_buildings.models import ResidentalComplex
from new_buildings.forms import NewBuildingsSearchForm
from feedback.models import Feedback
def corporation_benefit_plan(request):
return render(request, 'corporation_benefit_plan.html')
def index(request):
# Only 2 requests to DB
feedbacks = Feedback.objects.all(
)[:4].select_related().prefetch_related('social_media_links')
# Only 2 requests to DB
residental_complexes = ResidentalComplex.objects.filter(
is_popular=True).prefetch_related('type_of_complex')
context = {
'feedbacks': feedbacks,
'form': NewBuildingsSearchForm,
'residental_complexes': residental_complexes,
}
return render(request,
'index.html',
context,
)
def privacy_policy(request):
return render(request, 'privacy_policy.html')
def thanks(request):
return render(request, 'thanks.html')
| Use NewBuildingsSearchForm as main page search form. | Use NewBuildingsSearchForm as main page search form.
intead of non-complete SearchForm.
| Python | mit | Dybov/real_estate_agency,Dybov/real_estate_agency,Dybov/real_estate_agency | - from django.shortcuts import render, render_to_response
+ from django.shortcuts import render
- from django.template import RequestContext
- from new_buildings.models import Builder, ResidentalComplex, NewApartment
+ from new_buildings.models import ResidentalComplex
- from new_buildings.forms import SearchForm
+ from new_buildings.forms import NewBuildingsSearchForm
from feedback.models import Feedback
def corporation_benefit_plan(request):
return render(request, 'corporation_benefit_plan.html')
def index(request):
# Only 2 requests to DB
+ feedbacks = Feedback.objects.all(
- feedbacks = Feedback.objects.all()[:4].select_related().prefetch_related('social_media_links')
+ )[:4].select_related().prefetch_related('social_media_links')
# Only 2 requests to DB
residental_complexes = ResidentalComplex.objects.filter(
is_popular=True).prefetch_related('type_of_complex')
context = {
'feedbacks': feedbacks,
- 'form': SearchForm,
+ 'form': NewBuildingsSearchForm,
'residental_complexes': residental_complexes,
}
return render(request,
'index.html',
context,
)
def privacy_policy(request):
return render(request, 'privacy_policy.html')
def thanks(request):
return render(request, 'thanks.html')
| Use NewBuildingsSearchForm as main page search form. | ## Code Before:
from django.shortcuts import render, render_to_response
from django.template import RequestContext
from new_buildings.models import Builder, ResidentalComplex, NewApartment
from new_buildings.forms import SearchForm
from feedback.models import Feedback
def corporation_benefit_plan(request):
return render(request, 'corporation_benefit_plan.html')
def index(request):
# Only 2 requests to DB
feedbacks = Feedback.objects.all()[:4].select_related().prefetch_related('social_media_links')
# Only 2 requests to DB
residental_complexes = ResidentalComplex.objects.filter(
is_popular=True).prefetch_related('type_of_complex')
context = {
'feedbacks': feedbacks,
'form': SearchForm,
'residental_complexes': residental_complexes,
}
return render(request,
'index.html',
context,
)
def privacy_policy(request):
return render(request, 'privacy_policy.html')
def thanks(request):
return render(request, 'thanks.html')
## Instruction:
Use NewBuildingsSearchForm as main page search form.
## Code After:
from django.shortcuts import render
from new_buildings.models import ResidentalComplex
from new_buildings.forms import NewBuildingsSearchForm
from feedback.models import Feedback
def corporation_benefit_plan(request):
return render(request, 'corporation_benefit_plan.html')
def index(request):
# Only 2 requests to DB
feedbacks = Feedback.objects.all(
)[:4].select_related().prefetch_related('social_media_links')
# Only 2 requests to DB
residental_complexes = ResidentalComplex.objects.filter(
is_popular=True).prefetch_related('type_of_complex')
context = {
'feedbacks': feedbacks,
'form': NewBuildingsSearchForm,
'residental_complexes': residental_complexes,
}
return render(request,
'index.html',
context,
)
def privacy_policy(request):
return render(request, 'privacy_policy.html')
def thanks(request):
return render(request, 'thanks.html')
|
fe317230b6d2636b8a736c63be7769dd82663914 | libraries/SwitchManager.py | libraries/SwitchManager.py | class SwitchManager(object):
def extract_all_nodes(self, content):
return [e['node'] for e in content['nodeProperties']]
def extract_all_properties(self, content):
pass
| class SwitchManager(object):
def extract_all_nodes(self, content):
"""
Return all nodes.
"""
if isinstance(content,dict) or not content.has_key('nodeProperties'):
return None
else:
return [e.get('node') for e in content['nodeProperties']]
def extract_all_properties(self, content):
pass
| Add input check when getting all nodes. | Add input check when getting all nodes.
| Python | epl-1.0 | yeasy/robot_tool | class SwitchManager(object):
def extract_all_nodes(self, content):
+ """
+ Return all nodes.
+ """
+ if isinstance(content,dict) or not content.has_key('nodeProperties'):
+ return None
+ else:
- return [e['node'] for e in content['nodeProperties']]
+ return [e.get('node') for e in content['nodeProperties']]
def extract_all_properties(self, content):
pass
| Add input check when getting all nodes. | ## Code Before:
class SwitchManager(object):
def extract_all_nodes(self, content):
return [e['node'] for e in content['nodeProperties']]
def extract_all_properties(self, content):
pass
## Instruction:
Add input check when getting all nodes.
## Code After:
class SwitchManager(object):
def extract_all_nodes(self, content):
"""
Return all nodes.
"""
if isinstance(content,dict) or not content.has_key('nodeProperties'):
return None
else:
return [e.get('node') for e in content['nodeProperties']]
def extract_all_properties(self, content):
pass
|
99177cdc64bdec740557007800b610bff07ce46a | shivyc.py | shivyc.py |
import argparse
def get_arguments():
"""Set up the argument parser and return an object storing the
argument values.
return - An object storing argument values, as returned by
argparse.parse_args()
"""
parser = argparse.ArgumentParser(description="Compile C files.")
# The file name of the C file to compile. The file name gets saved to the
# file_name attribute of the returned object, but this parameter appears as
# "filename" (no underscore) on the command line.
parser.add_argument("file_name", metavar="filename")
return parser.parse_args()
def compile_code(source: str) -> str:
"""Compile the provided source code into assembly.
source - The C source code to compile.
return - The asm output
"""
return source
def main():
"""Load the input files, and dispatch to the compile function for the main
processing.
"""
arguments = get_arguments()
try:
c_file = open(arguments.file_name)
except IOError:
print("shivyc: error: no such file or directory: '{}'"
.format(arguments.file_name))
else:
compile_code(c_file.read())
c_file.close()
if __name__ == "__main__":
main()
|
import argparse
def get_arguments():
"""Set up the argument parser and return an object storing the
argument values.
return - An object storing argument values, as returned by
argparse.parse_args()
"""
parser = argparse.ArgumentParser(description="Compile C files.")
# The file name of the C file to compile. The file name gets saved to the
# file_name attribute of the returned object, but this parameter appears as
# "filename" (no underscore) on the command line.
parser.add_argument("file_name", metavar="filename")
return parser.parse_args()
def compile_code(source: str) -> str:
"""Compile the provided source code into assembly.
source - The C source code to compile.
return - The asm output
"""
return source
def main():
"""Load the input files and dispatch to the compile function for the main
processing.
The main function handles interfacing with the user, like reading the
command line arguments, printing errors, and generating output files. The
compilation logic is in the compile_code function to facilitate testing.
"""
arguments = get_arguments()
try:
c_file = open(arguments.file_name)
except IOError:
print("shivyc: error: no such file or directory: '{}'"
.format(arguments.file_name))
else:
compile_code(c_file.read())
c_file.close()
if __name__ == "__main__":
main()
| Improve commenting on main function | Improve commenting on main function
| Python | mit | ShivamSarodia/ShivyC,ShivamSarodia/ShivyC,ShivamSarodia/ShivyC |
import argparse
def get_arguments():
"""Set up the argument parser and return an object storing the
argument values.
return - An object storing argument values, as returned by
argparse.parse_args()
"""
parser = argparse.ArgumentParser(description="Compile C files.")
# The file name of the C file to compile. The file name gets saved to the
# file_name attribute of the returned object, but this parameter appears as
# "filename" (no underscore) on the command line.
parser.add_argument("file_name", metavar="filename")
return parser.parse_args()
def compile_code(source: str) -> str:
"""Compile the provided source code into assembly.
source - The C source code to compile.
return - The asm output
"""
return source
def main():
- """Load the input files, and dispatch to the compile function for the main
+ """Load the input files and dispatch to the compile function for the main
processing.
+
+ The main function handles interfacing with the user, like reading the
+ command line arguments, printing errors, and generating output files. The
+ compilation logic is in the compile_code function to facilitate testing.
"""
arguments = get_arguments()
try:
c_file = open(arguments.file_name)
except IOError:
print("shivyc: error: no such file or directory: '{}'"
.format(arguments.file_name))
else:
compile_code(c_file.read())
c_file.close()
if __name__ == "__main__":
main()
| Improve commenting on main function | ## Code Before:
import argparse
def get_arguments():
"""Set up the argument parser and return an object storing the
argument values.
return - An object storing argument values, as returned by
argparse.parse_args()
"""
parser = argparse.ArgumentParser(description="Compile C files.")
# The file name of the C file to compile. The file name gets saved to the
# file_name attribute of the returned object, but this parameter appears as
# "filename" (no underscore) on the command line.
parser.add_argument("file_name", metavar="filename")
return parser.parse_args()
def compile_code(source: str) -> str:
"""Compile the provided source code into assembly.
source - The C source code to compile.
return - The asm output
"""
return source
def main():
"""Load the input files, and dispatch to the compile function for the main
processing.
"""
arguments = get_arguments()
try:
c_file = open(arguments.file_name)
except IOError:
print("shivyc: error: no such file or directory: '{}'"
.format(arguments.file_name))
else:
compile_code(c_file.read())
c_file.close()
if __name__ == "__main__":
main()
## Instruction:
Improve commenting on main function
## Code After:
import argparse
def get_arguments():
"""Set up the argument parser and return an object storing the
argument values.
return - An object storing argument values, as returned by
argparse.parse_args()
"""
parser = argparse.ArgumentParser(description="Compile C files.")
# The file name of the C file to compile. The file name gets saved to the
# file_name attribute of the returned object, but this parameter appears as
# "filename" (no underscore) on the command line.
parser.add_argument("file_name", metavar="filename")
return parser.parse_args()
def compile_code(source: str) -> str:
"""Compile the provided source code into assembly.
source - The C source code to compile.
return - The asm output
"""
return source
def main():
"""Load the input files and dispatch to the compile function for the main
processing.
The main function handles interfacing with the user, like reading the
command line arguments, printing errors, and generating output files. The
compilation logic is in the compile_code function to facilitate testing.
"""
arguments = get_arguments()
try:
c_file = open(arguments.file_name)
except IOError:
print("shivyc: error: no such file or directory: '{}'"
.format(arguments.file_name))
else:
compile_code(c_file.read())
c_file.close()
if __name__ == "__main__":
main()
|
a34c9628c3f383e7b6f5eb521a9493f2b51d8811 | plata/reporting/views.py | plata/reporting/views.py | from decimal import Decimal
import StringIO
from django.contrib.admin.views.decorators import staff_member_required
from django.http import HttpResponse
from django.shortcuts import get_object_or_404
from pdfdocument.utils import pdf_response
import plata
import plata.reporting.product
import plata.reporting.order
@staff_member_required
def product_xls(request):
output = StringIO.StringIO()
workbook = plata.reporting.product.product_xls()
workbook.save(output)
response = HttpResponse(output.getvalue(), mimetype='application/vnd.ms-excel')
response['Content-Disposition'] = 'attachment; filename=products.xls'
return response
@staff_member_required
def order_pdf(request, order_id):
order = get_object_or_404(plata.shop_instance().order_model, pk=order_id)
order.shipping_cost = 8 / Decimal('1.076')
order.shipping_discount = 0
order.recalculate_total(save=False)
pdf, response = pdf_response('order-%09d' % order.id)
plata.reporting.order.order_pdf(pdf, order)
return response
| from decimal import Decimal
import StringIO
from django.contrib.admin.views.decorators import staff_member_required
from django.http import HttpResponse
from django.shortcuts import get_object_or_404
from pdfdocument.utils import pdf_response
import plata
import plata.reporting.product
import plata.reporting.order
@staff_member_required
def product_xls(request):
output = StringIO.StringIO()
workbook = plata.reporting.product.product_xls()
workbook.save(output)
response = HttpResponse(output.getvalue(), mimetype='application/vnd.ms-excel')
response['Content-Disposition'] = 'attachment; filename=products.xls'
return response
@staff_member_required
def order_pdf(request, order_id):
order = get_object_or_404(plata.shop_instance().order_model, pk=order_id)
pdf, response = pdf_response('order-%09d' % order.id)
plata.reporting.order.order_pdf(pdf, order)
return response
| Remove hardcoded shipping modification in order PDF view | Remove hardcoded shipping modification in order PDF view
| Python | bsd-3-clause | stefanklug/plata,armicron/plata,armicron/plata,allink/plata,armicron/plata | from decimal import Decimal
import StringIO
from django.contrib.admin.views.decorators import staff_member_required
from django.http import HttpResponse
from django.shortcuts import get_object_or_404
from pdfdocument.utils import pdf_response
import plata
import plata.reporting.product
import plata.reporting.order
@staff_member_required
def product_xls(request):
output = StringIO.StringIO()
workbook = plata.reporting.product.product_xls()
workbook.save(output)
response = HttpResponse(output.getvalue(), mimetype='application/vnd.ms-excel')
response['Content-Disposition'] = 'attachment; filename=products.xls'
return response
@staff_member_required
def order_pdf(request, order_id):
order = get_object_or_404(plata.shop_instance().order_model, pk=order_id)
- order.shipping_cost = 8 / Decimal('1.076')
- order.shipping_discount = 0
- order.recalculate_total(save=False)
-
pdf, response = pdf_response('order-%09d' % order.id)
plata.reporting.order.order_pdf(pdf, order)
return response
| Remove hardcoded shipping modification in order PDF view | ## Code Before:
from decimal import Decimal
import StringIO
from django.contrib.admin.views.decorators import staff_member_required
from django.http import HttpResponse
from django.shortcuts import get_object_or_404
from pdfdocument.utils import pdf_response
import plata
import plata.reporting.product
import plata.reporting.order
@staff_member_required
def product_xls(request):
output = StringIO.StringIO()
workbook = plata.reporting.product.product_xls()
workbook.save(output)
response = HttpResponse(output.getvalue(), mimetype='application/vnd.ms-excel')
response['Content-Disposition'] = 'attachment; filename=products.xls'
return response
@staff_member_required
def order_pdf(request, order_id):
order = get_object_or_404(plata.shop_instance().order_model, pk=order_id)
order.shipping_cost = 8 / Decimal('1.076')
order.shipping_discount = 0
order.recalculate_total(save=False)
pdf, response = pdf_response('order-%09d' % order.id)
plata.reporting.order.order_pdf(pdf, order)
return response
## Instruction:
Remove hardcoded shipping modification in order PDF view
## Code After:
from decimal import Decimal
import StringIO
from django.contrib.admin.views.decorators import staff_member_required
from django.http import HttpResponse
from django.shortcuts import get_object_or_404
from pdfdocument.utils import pdf_response
import plata
import plata.reporting.product
import plata.reporting.order
@staff_member_required
def product_xls(request):
output = StringIO.StringIO()
workbook = plata.reporting.product.product_xls()
workbook.save(output)
response = HttpResponse(output.getvalue(), mimetype='application/vnd.ms-excel')
response['Content-Disposition'] = 'attachment; filename=products.xls'
return response
@staff_member_required
def order_pdf(request, order_id):
order = get_object_or_404(plata.shop_instance().order_model, pk=order_id)
pdf, response = pdf_response('order-%09d' % order.id)
plata.reporting.order.order_pdf(pdf, order)
return response
|
7b1a0022b41dbf17de352e4686458e5250b28e49 | quantityfield/widgets.py | quantityfield/widgets.py | import re
from django.forms.widgets import MultiWidget, Select, NumberInput
from . import ureg
class QuantityWidget(MultiWidget):
def get_choices(self, allowed_types=None):
allowed_types = allowed_types or dir(ureg)
return [(x, x) for x in allowed_types]
def __init__(self, attrs=None, base_units=None, allowed_types=None):
choices = self.get_choices(allowed_types)
self.base_units = base_units
attrs = attrs or {}
attrs.setdefault('step', 'any')
widgets = (
NumberInput(attrs=attrs),
Select(attrs=attrs, choices=choices)
)
super(QuantityWidget, self).__init__(widgets, attrs)
def decompress(self, value):
non_decimal = re.compile(r'[^\d.]+')
if value:
number_value = non_decimal.sub('', str(value))
return [number_value, self.base_units]
return [None, self.base_units]
| import re
from django.forms.widgets import MultiWidget, Select, NumberInput
from . import ureg
class QuantityWidget(MultiWidget):
def get_choices(self, allowed_types=None):
allowed_types = allowed_types or dir(ureg)
return [(x, x) for x in allowed_types]
def __init__(self, attrs=None, base_units=None, allowed_types=None):
choices = self.get_choices(allowed_types)
self.base_units = base_units
attrs = attrs or {}
attrs.setdefault('step', 'any')
widgets = (
NumberInput(attrs=attrs),
Select(attrs=attrs, choices=choices)
)
super(QuantityWidget, self).__init__(widgets, attrs)
def decompress(self, value):
non_decimal = re.compile(r'[^\d.]+')
if value:
number_value = non_decimal.sub('', str(value))
return [number_value, self.base_units]
return [None, self.base_units]
| Fix indentation error from conversion to spaces | Fix indentation error from conversion to spaces
| Python | mit | bharling/django-pint,bharling/django-pint | import re
from django.forms.widgets import MultiWidget, Select, NumberInput
from . import ureg
class QuantityWidget(MultiWidget):
def get_choices(self, allowed_types=None):
allowed_types = allowed_types or dir(ureg)
return [(x, x) for x in allowed_types]
def __init__(self, attrs=None, base_units=None, allowed_types=None):
choices = self.get_choices(allowed_types)
self.base_units = base_units
attrs = attrs or {}
attrs.setdefault('step', 'any')
widgets = (
NumberInput(attrs=attrs),
Select(attrs=attrs, choices=choices)
)
super(QuantityWidget, self).__init__(widgets, attrs)
- def decompress(self, value):
+ def decompress(self, value):
- non_decimal = re.compile(r'[^\d.]+')
+ non_decimal = re.compile(r'[^\d.]+')
- if value:
+ if value:
- number_value = non_decimal.sub('', str(value))
+ number_value = non_decimal.sub('', str(value))
- return [number_value, self.base_units]
+ return [number_value, self.base_units]
- return [None, self.base_units]
+ return [None, self.base_units]
| Fix indentation error from conversion to spaces | ## Code Before:
import re
from django.forms.widgets import MultiWidget, Select, NumberInput
from . import ureg
class QuantityWidget(MultiWidget):
def get_choices(self, allowed_types=None):
allowed_types = allowed_types or dir(ureg)
return [(x, x) for x in allowed_types]
def __init__(self, attrs=None, base_units=None, allowed_types=None):
choices = self.get_choices(allowed_types)
self.base_units = base_units
attrs = attrs or {}
attrs.setdefault('step', 'any')
widgets = (
NumberInput(attrs=attrs),
Select(attrs=attrs, choices=choices)
)
super(QuantityWidget, self).__init__(widgets, attrs)
def decompress(self, value):
non_decimal = re.compile(r'[^\d.]+')
if value:
number_value = non_decimal.sub('', str(value))
return [number_value, self.base_units]
return [None, self.base_units]
## Instruction:
Fix indentation error from conversion to spaces
## Code After:
import re
from django.forms.widgets import MultiWidget, Select, NumberInput
from . import ureg
class QuantityWidget(MultiWidget):
def get_choices(self, allowed_types=None):
allowed_types = allowed_types or dir(ureg)
return [(x, x) for x in allowed_types]
def __init__(self, attrs=None, base_units=None, allowed_types=None):
choices = self.get_choices(allowed_types)
self.base_units = base_units
attrs = attrs or {}
attrs.setdefault('step', 'any')
widgets = (
NumberInput(attrs=attrs),
Select(attrs=attrs, choices=choices)
)
super(QuantityWidget, self).__init__(widgets, attrs)
def decompress(self, value):
non_decimal = re.compile(r'[^\d.]+')
if value:
number_value = non_decimal.sub('', str(value))
return [number_value, self.base_units]
return [None, self.base_units]
|
9ced61716167505875d3938ae01c08b61acc9392 | randterrainpy/terrain.py | randterrainpy/terrain.py | """This module is for the Terrain class, used for storing randomly generated terrain."""
class Terrain(object):
"""Container for a randomly generated area of terrain.
Attributes:
width (int): Width of generated terrain.
length (int): Length of generated terrain.
height_map (list): Map of heights of terrain. Values range from 0 to 1.
"""
def __init__(self, width, length):
"""Initializer for Terrain.
Args:
width (int): Width of terrain.
length (int): Height of terrain.
"""
self.width = width
self.length = length
self.height_map = [[0 for _ in self.width]] * self.length
def __getitem__(self, item):
"""Get an item at x-y coordinates.
Args:
item (tuple): 2-tuple of x and y coordinates.
Returns:
float: Height of terrain at coordinates, between 0 and 1.
"""
return self.height_map[item[1]][item[0]]
def __setitem__(self, key, value):
"""Set the height of an item.
Args:
key (tuple): 2-tuple of x and y coordinates.
value (float): New height of map at x and y coordinates, between 0 and 1.
"""
self.height_map[key[1]][key[0]] = value
| """This module is for the Terrain class, used for storing randomly generated terrain."""
class Terrain(object):
"""Container for a randomly generated area of terrain.
Attributes:
width (int): Width of generated terrain.
length (int): Length of generated terrain.
height_map (list): Map of heights of terrain. Values range from 0 to 1.
"""
def __init__(self, width, length):
"""Initializer for Terrain.
Args:
width (int): Width of terrain.
length (int): Height of terrain.
"""
self.width = width
self.length = length
self.height_map = [[0 for _ in self.width]] * self.length
def __getitem__(self, item):
"""Get an item at x-y coordinates.
Args:
item (tuple): 2-tuple of x and y coordinates.
Returns:
float: Height of terrain at coordinates, between 0 and 1.
"""
return self.height_map[item[1]][item[0]]
def __setitem__(self, key, value):
"""Set the height of an item.
Args:
key (tuple): 2-tuple of x and y coordinates.
value (float): New height of map at x and y coordinates, between 0 and 1.
"""
self.height_map[key[1]][key[0]] = value
def __add__(self, other):
"""Add two terrains, height by height.
Args:
other (Terrain): Other terrain to add self to. Must have same dimensions as self.
Returns:
Terrain: Terrain of self and other added together.
"""
result = Terrain(self.width, self.length)
for i in range(self.width):
for j in range(self.length):
result[i, j] = self[i, j] + other[i, j]
return result
| Add addition method to Terrain | Add addition method to Terrain
| Python | mit | jackromo/RandTerrainPy | """This module is for the Terrain class, used for storing randomly generated terrain."""
class Terrain(object):
"""Container for a randomly generated area of terrain.
Attributes:
width (int): Width of generated terrain.
length (int): Length of generated terrain.
height_map (list): Map of heights of terrain. Values range from 0 to 1.
"""
def __init__(self, width, length):
"""Initializer for Terrain.
Args:
width (int): Width of terrain.
length (int): Height of terrain.
"""
self.width = width
self.length = length
self.height_map = [[0 for _ in self.width]] * self.length
def __getitem__(self, item):
"""Get an item at x-y coordinates.
Args:
item (tuple): 2-tuple of x and y coordinates.
Returns:
float: Height of terrain at coordinates, between 0 and 1.
"""
return self.height_map[item[1]][item[0]]
def __setitem__(self, key, value):
"""Set the height of an item.
Args:
key (tuple): 2-tuple of x and y coordinates.
value (float): New height of map at x and y coordinates, between 0 and 1.
"""
self.height_map[key[1]][key[0]] = value
+ def __add__(self, other):
+ """Add two terrains, height by height.
+
+ Args:
+ other (Terrain): Other terrain to add self to. Must have same dimensions as self.
+
+ Returns:
+ Terrain: Terrain of self and other added together.
+
+ """
+ result = Terrain(self.width, self.length)
+ for i in range(self.width):
+ for j in range(self.length):
+ result[i, j] = self[i, j] + other[i, j]
+ return result
+ | Add addition method to Terrain | ## Code Before:
"""This module is for the Terrain class, used for storing randomly generated terrain."""
class Terrain(object):
"""Container for a randomly generated area of terrain.
Attributes:
width (int): Width of generated terrain.
length (int): Length of generated terrain.
height_map (list): Map of heights of terrain. Values range from 0 to 1.
"""
def __init__(self, width, length):
"""Initializer for Terrain.
Args:
width (int): Width of terrain.
length (int): Height of terrain.
"""
self.width = width
self.length = length
self.height_map = [[0 for _ in self.width]] * self.length
def __getitem__(self, item):
"""Get an item at x-y coordinates.
Args:
item (tuple): 2-tuple of x and y coordinates.
Returns:
float: Height of terrain at coordinates, between 0 and 1.
"""
return self.height_map[item[1]][item[0]]
def __setitem__(self, key, value):
"""Set the height of an item.
Args:
key (tuple): 2-tuple of x and y coordinates.
value (float): New height of map at x and y coordinates, between 0 and 1.
"""
self.height_map[key[1]][key[0]] = value
## Instruction:
Add addition method to Terrain
## Code After:
"""This module is for the Terrain class, used for storing randomly generated terrain."""
class Terrain(object):
"""Container for a randomly generated area of terrain.
Attributes:
width (int): Width of generated terrain.
length (int): Length of generated terrain.
height_map (list): Map of heights of terrain. Values range from 0 to 1.
"""
def __init__(self, width, length):
"""Initializer for Terrain.
Args:
width (int): Width of terrain.
length (int): Height of terrain.
"""
self.width = width
self.length = length
self.height_map = [[0 for _ in self.width]] * self.length
def __getitem__(self, item):
"""Get an item at x-y coordinates.
Args:
item (tuple): 2-tuple of x and y coordinates.
Returns:
float: Height of terrain at coordinates, between 0 and 1.
"""
return self.height_map[item[1]][item[0]]
def __setitem__(self, key, value):
"""Set the height of an item.
Args:
key (tuple): 2-tuple of x and y coordinates.
value (float): New height of map at x and y coordinates, between 0 and 1.
"""
self.height_map[key[1]][key[0]] = value
def __add__(self, other):
"""Add two terrains, height by height.
Args:
other (Terrain): Other terrain to add self to. Must have same dimensions as self.
Returns:
Terrain: Terrain of self and other added together.
"""
result = Terrain(self.width, self.length)
for i in range(self.width):
for j in range(self.length):
result[i, j] = self[i, j] + other[i, j]
return result
|
2e845dfd2695b1913f4603d88039049fa1eef923 | repositories.py | repositories.py | repositories = [
{
"owner": "talk-to",
"name": "Knock"
}
] | REPOSITORIES = [
{
"owner": "talk-to",
"name": "Knock"
}
] | Use capitalised name for constant | Use capitalised name for constant
| Python | mit | ayushgoel/LongShot | - repositories = [
+ REPOSITORIES = [
{
"owner": "talk-to",
"name": "Knock"
}
] | Use capitalised name for constant | ## Code Before:
repositories = [
{
"owner": "talk-to",
"name": "Knock"
}
]
## Instruction:
Use capitalised name for constant
## Code After:
REPOSITORIES = [
{
"owner": "talk-to",
"name": "Knock"
}
] |
aa7109d038a86f6a19a9fb4af96bd1199cd81330 | functest/opnfv_tests/openstack/snaps/snaps_utils.py | functest/opnfv_tests/openstack/snaps/snaps_utils.py |
from snaps.openstack.utils import neutron_utils, nova_utils
def get_ext_net_name(os_creds):
"""
Returns the first external network name
:param: os_creds: an instance of snaps OSCreds object
:return:
"""
neutron = neutron_utils.neutron_client(os_creds)
ext_nets = neutron_utils.get_external_networks(neutron)
return ext_nets[0].name if ext_nets else ""
def get_active_compute_cnt(os_creds):
"""
Returns the number of active compute servers
:param: os_creds: an instance of snaps OSCreds object
:return: the number of active compute servers
"""
nova = nova_utils.nova_client(os_creds)
computes = nova_utils.get_availability_zone_hosts(nova, zone_name='nova')
return len(computes)
|
from functest.utils.constants import CONST
from snaps.openstack.utils import neutron_utils, nova_utils
def get_ext_net_name(os_creds):
"""
Returns the configured external network name or
the first retrieved external network name
:param: os_creds: an instance of snaps OSCreds object
:return:
"""
neutron = neutron_utils.neutron_client(os_creds)
ext_nets = neutron_utils.get_external_networks(neutron)
if (hasattr(CONST, 'EXTERNAL_NETWORK')):
extnet_config = CONST.__getattribute__('EXTERNAL_NETWORK')
for ext_net in ext_nets:
if ext_net.name == extnet_config:
return extnet_config
return ext_nets[0].name if ext_nets else ""
def get_active_compute_cnt(os_creds):
"""
Returns the number of active compute servers
:param: os_creds: an instance of snaps OSCreds object
:return: the number of active compute servers
"""
nova = nova_utils.nova_client(os_creds)
computes = nova_utils.get_availability_zone_hosts(nova, zone_name='nova')
return len(computes)
| Support to specify the valid external network name | Support to specify the valid external network name
In some deployments, the retrieved external network by the
def get_external_networks in Snaps checked by "router:external"
is not available. So it is necessary to specify the available
external network as an env by user.
Change-Id: I333e91dd106ed307541a9a197280199fde86bd30
Signed-off-by: Linda Wang <81613bebe84fa394bbc7c5cc1c21989c9bff2c52@huawei.com>
| Python | apache-2.0 | opnfv/functest,mywulin/functest,opnfv/functest,mywulin/functest | +
+ from functest.utils.constants import CONST
from snaps.openstack.utils import neutron_utils, nova_utils
def get_ext_net_name(os_creds):
"""
- Returns the first external network name
+ Returns the configured external network name or
+ the first retrieved external network name
:param: os_creds: an instance of snaps OSCreds object
:return:
"""
neutron = neutron_utils.neutron_client(os_creds)
ext_nets = neutron_utils.get_external_networks(neutron)
+ if (hasattr(CONST, 'EXTERNAL_NETWORK')):
+ extnet_config = CONST.__getattribute__('EXTERNAL_NETWORK')
+ for ext_net in ext_nets:
+ if ext_net.name == extnet_config:
+ return extnet_config
return ext_nets[0].name if ext_nets else ""
def get_active_compute_cnt(os_creds):
"""
Returns the number of active compute servers
:param: os_creds: an instance of snaps OSCreds object
:return: the number of active compute servers
"""
nova = nova_utils.nova_client(os_creds)
computes = nova_utils.get_availability_zone_hosts(nova, zone_name='nova')
return len(computes)
| Support to specify the valid external network name | ## Code Before:
from snaps.openstack.utils import neutron_utils, nova_utils
def get_ext_net_name(os_creds):
"""
Returns the first external network name
:param: os_creds: an instance of snaps OSCreds object
:return:
"""
neutron = neutron_utils.neutron_client(os_creds)
ext_nets = neutron_utils.get_external_networks(neutron)
return ext_nets[0].name if ext_nets else ""
def get_active_compute_cnt(os_creds):
"""
Returns the number of active compute servers
:param: os_creds: an instance of snaps OSCreds object
:return: the number of active compute servers
"""
nova = nova_utils.nova_client(os_creds)
computes = nova_utils.get_availability_zone_hosts(nova, zone_name='nova')
return len(computes)
## Instruction:
Support to specify the valid external network name
## Code After:
from functest.utils.constants import CONST
from snaps.openstack.utils import neutron_utils, nova_utils
def get_ext_net_name(os_creds):
"""
Returns the configured external network name or
the first retrieved external network name
:param: os_creds: an instance of snaps OSCreds object
:return:
"""
neutron = neutron_utils.neutron_client(os_creds)
ext_nets = neutron_utils.get_external_networks(neutron)
if (hasattr(CONST, 'EXTERNAL_NETWORK')):
extnet_config = CONST.__getattribute__('EXTERNAL_NETWORK')
for ext_net in ext_nets:
if ext_net.name == extnet_config:
return extnet_config
return ext_nets[0].name if ext_nets else ""
def get_active_compute_cnt(os_creds):
"""
Returns the number of active compute servers
:param: os_creds: an instance of snaps OSCreds object
:return: the number of active compute servers
"""
nova = nova_utils.nova_client(os_creds)
computes = nova_utils.get_availability_zone_hosts(nova, zone_name='nova')
return len(computes)
|
acd4238dce39464e99964227dca7758cffedca39 | gaphor/UML/classes/tests/test_containmentconnect.py | gaphor/UML/classes/tests/test_containmentconnect.py | """Test connection of containment relationship."""
from gaphor import UML
from gaphor.diagram.tests.fixtures import allow, connect
from gaphor.UML.classes import PackageItem
from gaphor.UML.classes.containment import ContainmentItem
def test_containment_package_glue(create):
"""Test containment glue to two package items."""
pkg1 = create(PackageItem, UML.Package)
pkg2 = create(PackageItem, UML.Package)
containment = create(ContainmentItem)
glued = allow(containment, containment.head, pkg1)
assert glued
connect(containment, containment.head, pkg1)
glued = allow(containment, containment.tail, pkg2)
assert glued
| """Test connection of containment relationship."""
from gaphor import UML
from gaphor.diagram.tests.fixtures import allow, connect
from gaphor.UML.classes import ClassItem, PackageItem
from gaphor.UML.classes.containment import ContainmentItem
def test_containment_package_glue(create):
"""Test containment glue to two package items."""
pkg1 = create(PackageItem, UML.Package)
pkg2 = create(PackageItem, UML.Package)
containment = create(ContainmentItem)
glued = allow(containment, containment.head, pkg1)
assert glued
connect(containment, containment.head, pkg1)
glued = allow(containment, containment.tail, pkg2)
assert glued
def test_containment_package_class(create, diagram):
"""Test containment connecting to a package and a class."""
package = create(ContainmentItem, UML.Package)
line = create(ContainmentItem)
ac = create(ClassItem, UML.Class)
connect(line, line.head, package)
connect(line, line.tail, ac)
assert diagram.connections.get_connection(line.tail).connected is ac
assert len(package.subject.ownedElement) == 1
assert ac.subject in package.subject.ownedElement
| Add test for connecting containment to package and a class | Add test for connecting containment to package and a class [skip ci]
Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>
| Python | lgpl-2.1 | amolenaar/gaphor,amolenaar/gaphor | """Test connection of containment relationship."""
from gaphor import UML
from gaphor.diagram.tests.fixtures import allow, connect
- from gaphor.UML.classes import PackageItem
+ from gaphor.UML.classes import ClassItem, PackageItem
from gaphor.UML.classes.containment import ContainmentItem
def test_containment_package_glue(create):
"""Test containment glue to two package items."""
pkg1 = create(PackageItem, UML.Package)
pkg2 = create(PackageItem, UML.Package)
containment = create(ContainmentItem)
glued = allow(containment, containment.head, pkg1)
assert glued
connect(containment, containment.head, pkg1)
glued = allow(containment, containment.tail, pkg2)
assert glued
+
+ def test_containment_package_class(create, diagram):
+ """Test containment connecting to a package and a class."""
+ package = create(ContainmentItem, UML.Package)
+ line = create(ContainmentItem)
+ ac = create(ClassItem, UML.Class)
+
+ connect(line, line.head, package)
+ connect(line, line.tail, ac)
+ assert diagram.connections.get_connection(line.tail).connected is ac
+ assert len(package.subject.ownedElement) == 1
+ assert ac.subject in package.subject.ownedElement
+ | Add test for connecting containment to package and a class | ## Code Before:
"""Test connection of containment relationship."""
from gaphor import UML
from gaphor.diagram.tests.fixtures import allow, connect
from gaphor.UML.classes import PackageItem
from gaphor.UML.classes.containment import ContainmentItem
def test_containment_package_glue(create):
"""Test containment glue to two package items."""
pkg1 = create(PackageItem, UML.Package)
pkg2 = create(PackageItem, UML.Package)
containment = create(ContainmentItem)
glued = allow(containment, containment.head, pkg1)
assert glued
connect(containment, containment.head, pkg1)
glued = allow(containment, containment.tail, pkg2)
assert glued
## Instruction:
Add test for connecting containment to package and a class
## Code After:
"""Test connection of containment relationship."""
from gaphor import UML
from gaphor.diagram.tests.fixtures import allow, connect
from gaphor.UML.classes import ClassItem, PackageItem
from gaphor.UML.classes.containment import ContainmentItem
def test_containment_package_glue(create):
"""Test containment glue to two package items."""
pkg1 = create(PackageItem, UML.Package)
pkg2 = create(PackageItem, UML.Package)
containment = create(ContainmentItem)
glued = allow(containment, containment.head, pkg1)
assert glued
connect(containment, containment.head, pkg1)
glued = allow(containment, containment.tail, pkg2)
assert glued
def test_containment_package_class(create, diagram):
"""Test containment connecting to a package and a class."""
package = create(ContainmentItem, UML.Package)
line = create(ContainmentItem)
ac = create(ClassItem, UML.Class)
connect(line, line.head, package)
connect(line, line.tail, ac)
assert diagram.connections.get_connection(line.tail).connected is ac
assert len(package.subject.ownedElement) == 1
assert ac.subject in package.subject.ownedElement
|
413ba364dc35a7186953d02bb7cc8cf705371873 | contentious/constants.py | contentious/constants.py | from django.conf import settings
SELF_CLOSING_HTML_TAGS = getattr(settings,
'CONTENTIOUS_SELF_CLOSING_HTML_TAGS', ['img', 'br', 'hr', 'meta'])
#Note, the Javascript plugin has its own seprate copy of this:
TREAT_CONTENT_AS_HTML_TAGS = getattr(settings,
'CONTENTIOUS_TREAT_CONTENT_AS_HTML_TAGS', ['div', 'select', 'ul'])
| from django.conf import settings
SELF_CLOSING_HTML_TAGS = ['img', 'br', 'hr', 'meta']
#Note, the Javascript plugin has its own seprate copy of this:
TREAT_CONTENT_AS_HTML_TAGS = getattr(settings,
'CONTENTIOUS_TREAT_CONTENT_AS_HTML_TAGS', ['div', 'select', 'ul'])
| Remove SELF_CLOSING_HTML_TAGS as a configurable option | Remove SELF_CLOSING_HTML_TAGS as a configurable option
| Python | bsd-2-clause | potatolondon/contentious,potatolondon/contentious | from django.conf import settings
- SELF_CLOSING_HTML_TAGS = getattr(settings,
- 'CONTENTIOUS_SELF_CLOSING_HTML_TAGS', ['img', 'br', 'hr', 'meta'])
+ SELF_CLOSING_HTML_TAGS = ['img', 'br', 'hr', 'meta']
#Note, the Javascript plugin has its own seprate copy of this:
TREAT_CONTENT_AS_HTML_TAGS = getattr(settings,
'CONTENTIOUS_TREAT_CONTENT_AS_HTML_TAGS', ['div', 'select', 'ul'])
| Remove SELF_CLOSING_HTML_TAGS as a configurable option | ## Code Before:
from django.conf import settings
SELF_CLOSING_HTML_TAGS = getattr(settings,
'CONTENTIOUS_SELF_CLOSING_HTML_TAGS', ['img', 'br', 'hr', 'meta'])
#Note, the Javascript plugin has its own seprate copy of this:
TREAT_CONTENT_AS_HTML_TAGS = getattr(settings,
'CONTENTIOUS_TREAT_CONTENT_AS_HTML_TAGS', ['div', 'select', 'ul'])
## Instruction:
Remove SELF_CLOSING_HTML_TAGS as a configurable option
## Code After:
from django.conf import settings
SELF_CLOSING_HTML_TAGS = ['img', 'br', 'hr', 'meta']
#Note, the Javascript plugin has its own seprate copy of this:
TREAT_CONTENT_AS_HTML_TAGS = getattr(settings,
'CONTENTIOUS_TREAT_CONTENT_AS_HTML_TAGS', ['div', 'select', 'ul'])
|
182cd3b73382bb150111198e5fcbfa43a6bd416f | cbagent/collectors/libstats/typeperfstats.py | cbagent/collectors/libstats/typeperfstats.py | from cbagent.collectors.libstats.remotestats import RemoteStats, parallel_task
class TPStats(RemoteStats):
METRICS = (
("rss", 1), # already in bytes
)
def __init__(self, hosts, workers, user, password):
super().__init__(hosts, workers, user, password)
self.typeperf_cmd = "typeperf \"\\Process(*{}*)\\Working Set\" -sc 1|sed '3q;d'"
@parallel_task(server_side=True)
def get_samples(self, process):
samples = {}
if process == "beam.smp":
stdout = self.run(self.typeperf_cmd.format("erl"))
values = stdout.split(',')[1:5]
elif process == "memcached":
stdout = self.run(self.typeperf_cmd.format(process))
values = stdout.split(',')[1:2]
else:
return samples
sum_rss = 0
if stdout:
for v in values:
v = float(v.replace('"', ''))
sum_rss += v
metric, multiplier = self.METRICS[0]
title = "{}_{}".format(process, metric)
samples[title] = float(sum_rss) * multiplier
return samples
| from cbagent.collectors.libstats.remotestats import RemoteStats, parallel_task
class TPStats(RemoteStats):
METRICS = (
("rss", 1), # already in bytes
)
def __init__(self, hosts, workers, user, password):
super().__init__(hosts, workers, user, password)
self.typeperf_cmd = "typeperf \"\\Process(*{}*)\\Working Set\" -sc 1|sed '3q;d'"
@parallel_task(server_side=True)
def get_server_samples(self, process):
samples = {}
if process == "beam.smp":
stdout = self.run(self.typeperf_cmd.format("erl"))
values = stdout.split(',')[1:5]
elif process == "memcached":
stdout = self.run(self.typeperf_cmd.format(process))
values = stdout.split(',')[1:2]
else:
return samples
sum_rss = 0
if stdout:
for v in values:
v = float(v.replace('"', ''))
sum_rss += v
metric, multiplier = self.METRICS[0]
title = "{}_{}".format(process, metric)
samples[title] = float(sum_rss) * multiplier
return samples
def get_client_samples(self, process):
pass
| Add missing methods to TPStats | Add missing methods to TPStats
Change-Id: I332a83f3816ee30597288180ed344da3161861f8
Reviewed-on: http://review.couchbase.org/79675
Tested-by: Build Bot <80754af91bfb6d1073585b046fe0a474ce868509@couchbase.com>
Reviewed-by: Pavel Paulau <dd88eded64e90046a680e3a6c0828ceb8fe8a0e7@gmail.com>
| Python | apache-2.0 | pavel-paulau/perfrunner,pavel-paulau/perfrunner,pavel-paulau/perfrunner,couchbase/perfrunner,couchbase/perfrunner,couchbase/perfrunner,pavel-paulau/perfrunner,couchbase/perfrunner,pavel-paulau/perfrunner,couchbase/perfrunner,couchbase/perfrunner | from cbagent.collectors.libstats.remotestats import RemoteStats, parallel_task
class TPStats(RemoteStats):
METRICS = (
("rss", 1), # already in bytes
)
def __init__(self, hosts, workers, user, password):
super().__init__(hosts, workers, user, password)
self.typeperf_cmd = "typeperf \"\\Process(*{}*)\\Working Set\" -sc 1|sed '3q;d'"
@parallel_task(server_side=True)
- def get_samples(self, process):
+ def get_server_samples(self, process):
samples = {}
if process == "beam.smp":
stdout = self.run(self.typeperf_cmd.format("erl"))
values = stdout.split(',')[1:5]
elif process == "memcached":
stdout = self.run(self.typeperf_cmd.format(process))
values = stdout.split(',')[1:2]
else:
return samples
sum_rss = 0
if stdout:
for v in values:
v = float(v.replace('"', ''))
sum_rss += v
metric, multiplier = self.METRICS[0]
title = "{}_{}".format(process, metric)
samples[title] = float(sum_rss) * multiplier
return samples
+ def get_client_samples(self, process):
+ pass
+ | Add missing methods to TPStats | ## Code Before:
from cbagent.collectors.libstats.remotestats import RemoteStats, parallel_task
class TPStats(RemoteStats):
METRICS = (
("rss", 1), # already in bytes
)
def __init__(self, hosts, workers, user, password):
super().__init__(hosts, workers, user, password)
self.typeperf_cmd = "typeperf \"\\Process(*{}*)\\Working Set\" -sc 1|sed '3q;d'"
@parallel_task(server_side=True)
def get_samples(self, process):
samples = {}
if process == "beam.smp":
stdout = self.run(self.typeperf_cmd.format("erl"))
values = stdout.split(',')[1:5]
elif process == "memcached":
stdout = self.run(self.typeperf_cmd.format(process))
values = stdout.split(',')[1:2]
else:
return samples
sum_rss = 0
if stdout:
for v in values:
v = float(v.replace('"', ''))
sum_rss += v
metric, multiplier = self.METRICS[0]
title = "{}_{}".format(process, metric)
samples[title] = float(sum_rss) * multiplier
return samples
## Instruction:
Add missing methods to TPStats
## Code After:
from cbagent.collectors.libstats.remotestats import RemoteStats, parallel_task
class TPStats(RemoteStats):
METRICS = (
("rss", 1), # already in bytes
)
def __init__(self, hosts, workers, user, password):
super().__init__(hosts, workers, user, password)
self.typeperf_cmd = "typeperf \"\\Process(*{}*)\\Working Set\" -sc 1|sed '3q;d'"
@parallel_task(server_side=True)
def get_server_samples(self, process):
samples = {}
if process == "beam.smp":
stdout = self.run(self.typeperf_cmd.format("erl"))
values = stdout.split(',')[1:5]
elif process == "memcached":
stdout = self.run(self.typeperf_cmd.format(process))
values = stdout.split(',')[1:2]
else:
return samples
sum_rss = 0
if stdout:
for v in values:
v = float(v.replace('"', ''))
sum_rss += v
metric, multiplier = self.METRICS[0]
title = "{}_{}".format(process, metric)
samples[title] = float(sum_rss) * multiplier
return samples
def get_client_samples(self, process):
pass
|
c45fc698da9783b561cca69363ec4998622e9ac0 | mint/rest/db/capsulemgr.py | mint/rest/db/capsulemgr.py |
from conary.lib import util
from mint.rest.db import manager
import rpath_capsule_indexer
class CapsuleManager(manager.Manager):
def getIndexerConfig(self):
capsuleDataDir = util.joinPaths(self.cfg.dataPath, 'capsules')
cfg = rpath_capsule_indexer.IndexerConfig()
cfg.configLine("store sqlite:///%s/database.sqlite" %
capsuleDataDir)
cfg.configLine("indexDir %s/packages" % capsuleDataDir)
cfg.configLine("systemsPath %s/systems" % capsuleDataDir)
dataSources = self.db.platformMgr.listPlatformSources().platformSource
# XXX we only deal with RHN for now
if dataSources:
cfg.configLine("user RHN %s %s" % (dataSources[0].username,
dataSources[0].password))
# XXX channels are hardcoded for now
cfg.configLine("channels rhel-i386-as-4")
cfg.configLine("channels rhel-x86_64-as-4")
cfg.configLine("channels rhel-i386-server-5")
cfg.configLine("channels rhel-x86_64-server-5")
util.mkdirChain(capsuleDataDir)
return cfg
def getIndexer(self):
cfg = self.getIndexerConfig()
return rpath_capsule_indexer.Indexer(cfg)
|
from conary.lib import util
from mint.rest.db import manager
import rpath_capsule_indexer
class CapsuleManager(manager.Manager):
def getIndexerConfig(self):
capsuleDataDir = util.joinPaths(self.cfg.dataPath, 'capsules')
cfg = rpath_capsule_indexer.IndexerConfig()
dbDriver = self.db.db.driver
dbConnectString = self.db.db.db.database
cfg.configLine("store %s:///%s" % (dbDriver, dbConnectString))
cfg.configLine("indexDir %s/packages" % capsuleDataDir)
cfg.configLine("systemsPath %s/systems" % capsuleDataDir)
dataSources = self.db.platformMgr.listPlatformSources().platformSource
# XXX we only deal with RHN for now
if dataSources:
cfg.configLine("user RHN %s %s" % (dataSources[0].username,
dataSources[0].password))
# XXX channels are hardcoded for now
cfg.configLine("channels rhel-i386-as-4")
cfg.configLine("channels rhel-x86_64-as-4")
cfg.configLine("channels rhel-i386-server-5")
cfg.configLine("channels rhel-x86_64-server-5")
util.mkdirChain(capsuleDataDir)
return cfg
def getIndexer(self):
cfg = self.getIndexerConfig()
return rpath_capsule_indexer.Indexer(cfg)
| Use the mint database for capsule data | Use the mint database for capsule data
| Python | apache-2.0 | sassoftware/mint,sassoftware/mint,sassoftware/mint,sassoftware/mint,sassoftware/mint |
from conary.lib import util
from mint.rest.db import manager
import rpath_capsule_indexer
class CapsuleManager(manager.Manager):
def getIndexerConfig(self):
capsuleDataDir = util.joinPaths(self.cfg.dataPath, 'capsules')
cfg = rpath_capsule_indexer.IndexerConfig()
- cfg.configLine("store sqlite:///%s/database.sqlite" %
- capsuleDataDir)
+ dbDriver = self.db.db.driver
+ dbConnectString = self.db.db.db.database
+ cfg.configLine("store %s:///%s" % (dbDriver, dbConnectString))
cfg.configLine("indexDir %s/packages" % capsuleDataDir)
cfg.configLine("systemsPath %s/systems" % capsuleDataDir)
dataSources = self.db.platformMgr.listPlatformSources().platformSource
# XXX we only deal with RHN for now
if dataSources:
cfg.configLine("user RHN %s %s" % (dataSources[0].username,
dataSources[0].password))
# XXX channels are hardcoded for now
cfg.configLine("channels rhel-i386-as-4")
cfg.configLine("channels rhel-x86_64-as-4")
cfg.configLine("channels rhel-i386-server-5")
cfg.configLine("channels rhel-x86_64-server-5")
util.mkdirChain(capsuleDataDir)
return cfg
def getIndexer(self):
cfg = self.getIndexerConfig()
return rpath_capsule_indexer.Indexer(cfg)
| Use the mint database for capsule data | ## Code Before:
from conary.lib import util
from mint.rest.db import manager
import rpath_capsule_indexer
class CapsuleManager(manager.Manager):
def getIndexerConfig(self):
capsuleDataDir = util.joinPaths(self.cfg.dataPath, 'capsules')
cfg = rpath_capsule_indexer.IndexerConfig()
cfg.configLine("store sqlite:///%s/database.sqlite" %
capsuleDataDir)
cfg.configLine("indexDir %s/packages" % capsuleDataDir)
cfg.configLine("systemsPath %s/systems" % capsuleDataDir)
dataSources = self.db.platformMgr.listPlatformSources().platformSource
# XXX we only deal with RHN for now
if dataSources:
cfg.configLine("user RHN %s %s" % (dataSources[0].username,
dataSources[0].password))
# XXX channels are hardcoded for now
cfg.configLine("channels rhel-i386-as-4")
cfg.configLine("channels rhel-x86_64-as-4")
cfg.configLine("channels rhel-i386-server-5")
cfg.configLine("channels rhel-x86_64-server-5")
util.mkdirChain(capsuleDataDir)
return cfg
def getIndexer(self):
cfg = self.getIndexerConfig()
return rpath_capsule_indexer.Indexer(cfg)
## Instruction:
Use the mint database for capsule data
## Code After:
from conary.lib import util
from mint.rest.db import manager
import rpath_capsule_indexer
class CapsuleManager(manager.Manager):
def getIndexerConfig(self):
capsuleDataDir = util.joinPaths(self.cfg.dataPath, 'capsules')
cfg = rpath_capsule_indexer.IndexerConfig()
dbDriver = self.db.db.driver
dbConnectString = self.db.db.db.database
cfg.configLine("store %s:///%s" % (dbDriver, dbConnectString))
cfg.configLine("indexDir %s/packages" % capsuleDataDir)
cfg.configLine("systemsPath %s/systems" % capsuleDataDir)
dataSources = self.db.platformMgr.listPlatformSources().platformSource
# XXX we only deal with RHN for now
if dataSources:
cfg.configLine("user RHN %s %s" % (dataSources[0].username,
dataSources[0].password))
# XXX channels are hardcoded for now
cfg.configLine("channels rhel-i386-as-4")
cfg.configLine("channels rhel-x86_64-as-4")
cfg.configLine("channels rhel-i386-server-5")
cfg.configLine("channels rhel-x86_64-server-5")
util.mkdirChain(capsuleDataDir)
return cfg
def getIndexer(self):
cfg = self.getIndexerConfig()
return rpath_capsule_indexer.Indexer(cfg)
|
21059428d95c27cf043ada2e299a4cf3982a4233 | python/printbag.py | python/printbag.py |
"""LIDAR datatype format is:
(
timestamp (long),
flag (bool saved as int),
accelerometer[3] (double),
gps[3] (double),
distance[LIDAR_NUM_ANGLES] (long),
)
'int' and 'long' are the same size on the raspberry pi (32 bits).
"""
import sys
import rosbag
def decode_bag(bag):
topics = ['/scan', '/flagbutton_pressed']
return [message for message in bag.read_messages(topics=topics)]
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} <rosbag> [<outfile>] \n\n'
'Print contents of rosbag file. If <outfile> is provided, \n'
'write contents of rosbag file to <outfile> in the legacy \n'
'lidar binary format.').format(__file__))
sys.exit(1)
outfile = None
filename = sys.argv[1]
if len(sys.argv) == 3:
outfile = sys.argv[2]
with rosbag.Bag(filename) as bag:
print(decode_bag(bag))
sys.exit()
|
"""LIDAR datatype format is:
(
timestamp (long),
flag (bool saved as int),
accelerometer[3] (double),
gps[3] (double),
distance[LIDAR_NUM_ANGLES] (long),
)
'int' and 'long' are the same size on the raspberry pi (32 bits).
"""
import sys
import rosbag
def print_bag(bag):
topics = ['/scan', '/flagbutton_pressed']
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} <rosbag> [<outfile>] \n\n'
'Print contents of rosbag file. If <outfile> is provided, \n'
'write contents of rosbag file to <outfile> in the legacy \n'
'lidar binary format.').format(__file__))
sys.exit(1)
outfile = None
filename = sys.argv[1]
if len(sys.argv) == 3:
outfile = sys.argv[2]
with rosbag.Bag(filename) as bag:
print_bag(bag)
sys.exit()
| Print out bag contents for lidar and button topics | Print out bag contents for lidar and button topics
| Python | bsd-2-clause | oliverlee/antlia |
"""LIDAR datatype format is:
(
timestamp (long),
flag (bool saved as int),
accelerometer[3] (double),
gps[3] (double),
distance[LIDAR_NUM_ANGLES] (long),
)
'int' and 'long' are the same size on the raspberry pi (32 bits).
"""
import sys
import rosbag
- def decode_bag(bag):
+ def print_bag(bag):
topics = ['/scan', '/flagbutton_pressed']
- return [message for message in bag.read_messages(topics=topics)]
+ for message in bag.read_messages(topics=topics):
+ print(message)
+
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} <rosbag> [<outfile>] \n\n'
'Print contents of rosbag file. If <outfile> is provided, \n'
'write contents of rosbag file to <outfile> in the legacy \n'
'lidar binary format.').format(__file__))
sys.exit(1)
outfile = None
filename = sys.argv[1]
if len(sys.argv) == 3:
outfile = sys.argv[2]
with rosbag.Bag(filename) as bag:
- print(decode_bag(bag))
+ print_bag(bag)
sys.exit()
| Print out bag contents for lidar and button topics | ## Code Before:
"""LIDAR datatype format is:
(
timestamp (long),
flag (bool saved as int),
accelerometer[3] (double),
gps[3] (double),
distance[LIDAR_NUM_ANGLES] (long),
)
'int' and 'long' are the same size on the raspberry pi (32 bits).
"""
import sys
import rosbag
def decode_bag(bag):
topics = ['/scan', '/flagbutton_pressed']
return [message for message in bag.read_messages(topics=topics)]
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} <rosbag> [<outfile>] \n\n'
'Print contents of rosbag file. If <outfile> is provided, \n'
'write contents of rosbag file to <outfile> in the legacy \n'
'lidar binary format.').format(__file__))
sys.exit(1)
outfile = None
filename = sys.argv[1]
if len(sys.argv) == 3:
outfile = sys.argv[2]
with rosbag.Bag(filename) as bag:
print(decode_bag(bag))
sys.exit()
## Instruction:
Print out bag contents for lidar and button topics
## Code After:
"""LIDAR datatype format is:
(
timestamp (long),
flag (bool saved as int),
accelerometer[3] (double),
gps[3] (double),
distance[LIDAR_NUM_ANGLES] (long),
)
'int' and 'long' are the same size on the raspberry pi (32 bits).
"""
import sys
import rosbag
def print_bag(bag):
topics = ['/scan', '/flagbutton_pressed']
for message in bag.read_messages(topics=topics):
print(message)
if __name__ == '__main__':
if len(sys.argv) < 2:
print(('Usage: {} <rosbag> [<outfile>] \n\n'
'Print contents of rosbag file. If <outfile> is provided, \n'
'write contents of rosbag file to <outfile> in the legacy \n'
'lidar binary format.').format(__file__))
sys.exit(1)
outfile = None
filename = sys.argv[1]
if len(sys.argv) == 3:
outfile = sys.argv[2]
with rosbag.Bag(filename) as bag:
print_bag(bag)
sys.exit()
|
78cca16df6a5cdd90ec92e64455215c4b7292fae | report_coverage.py | report_coverage.py | import json
import os
import sys
from coveralls import Coveralls, cli
# Patch coveralls to get javascript coverage from mocha
orig_get_coverage = Coveralls.get_coverage
def get_coverage_with_js(self):
report = orig_get_coverage(self)
js_files = json.load(open('.coverage-js'))['files']
js_report = []
for f in js_files:
source = '\n'.join(open(f['filename']).readlines())
name = os.path.relpath(f['filename'])
coverage = []
for v in f['source'].values():
coverage.append(v['coverage'] if v['coverage'] != '' else None)
js_report.append({
'source': source,
'name': name,
'coverage': coverage}
)
report += js_report
return report
Coveralls.get_coverage = get_coverage_with_js
cli.main(sys.argv[1:])
| import json
import os
import sys
from coveralls import Coveralls, cli
# Patch coveralls to get javascript coverage from mocha
orig_get_coverage = Coveralls.get_coverage
def get_coverage_with_js(self):
report = orig_get_coverage(self)
js_files = json.load(open('.coverage-js'))['files']
js_report = []
for f in js_files:
source = '\n'.join(open(f['filename']).readlines())
name = os.path.relpath(f['filename'])
coverage = []
# Create sorted coverage array from original dict
for k, v in sorted(f['source'].items(), key=lambda x:int(x[0])):
coverage.append(v['coverage'] if v['coverage'] != '' else None)
js_report.append({
'source': source,
'name': name,
'coverage': coverage}
)
report += js_report
return report
Coveralls.get_coverage = get_coverage_with_js
cli.main(sys.argv[1:])
| Sort line coverage info when reporting | Sort line coverage info when reporting
| Python | apache-2.0 | exekias/django-achilles,exekias/django-achilles | import json
import os
import sys
from coveralls import Coveralls, cli
# Patch coveralls to get javascript coverage from mocha
orig_get_coverage = Coveralls.get_coverage
def get_coverage_with_js(self):
report = orig_get_coverage(self)
js_files = json.load(open('.coverage-js'))['files']
js_report = []
for f in js_files:
source = '\n'.join(open(f['filename']).readlines())
name = os.path.relpath(f['filename'])
coverage = []
- for v in f['source'].values():
+
+ # Create sorted coverage array from original dict
+ for k, v in sorted(f['source'].items(), key=lambda x:int(x[0])):
coverage.append(v['coverage'] if v['coverage'] != '' else None)
js_report.append({
'source': source,
'name': name,
'coverage': coverage}
)
report += js_report
return report
Coveralls.get_coverage = get_coverage_with_js
cli.main(sys.argv[1:])
| Sort line coverage info when reporting | ## Code Before:
import json
import os
import sys
from coveralls import Coveralls, cli
# Patch coveralls to get javascript coverage from mocha
orig_get_coverage = Coveralls.get_coverage
def get_coverage_with_js(self):
report = orig_get_coverage(self)
js_files = json.load(open('.coverage-js'))['files']
js_report = []
for f in js_files:
source = '\n'.join(open(f['filename']).readlines())
name = os.path.relpath(f['filename'])
coverage = []
for v in f['source'].values():
coverage.append(v['coverage'] if v['coverage'] != '' else None)
js_report.append({
'source': source,
'name': name,
'coverage': coverage}
)
report += js_report
return report
Coveralls.get_coverage = get_coverage_with_js
cli.main(sys.argv[1:])
## Instruction:
Sort line coverage info when reporting
## Code After:
import json
import os
import sys
from coveralls import Coveralls, cli
# Patch coveralls to get javascript coverage from mocha
orig_get_coverage = Coveralls.get_coverage
def get_coverage_with_js(self):
report = orig_get_coverage(self)
js_files = json.load(open('.coverage-js'))['files']
js_report = []
for f in js_files:
source = '\n'.join(open(f['filename']).readlines())
name = os.path.relpath(f['filename'])
coverage = []
# Create sorted coverage array from original dict
for k, v in sorted(f['source'].items(), key=lambda x:int(x[0])):
coverage.append(v['coverage'] if v['coverage'] != '' else None)
js_report.append({
'source': source,
'name': name,
'coverage': coverage}
)
report += js_report
return report
Coveralls.get_coverage = get_coverage_with_js
cli.main(sys.argv[1:])
|
4ae8302a3d91ca1e9601e0c51cb58a69f1c08cb5 | setup.py | setup.py | """bibpy module setup script for distribution."""
from __future__ import with_statement
import os
import distutils.core
def get_version(filename):
with open(filename) as fh:
for line in fh:
if line.startswith('__version__'):
return line.split('=')[-1].strip()[1:-1]
distutils.core.setup(
name='bibpy',
version=get_version(os.path.join('bibpy', '__init__.py')),
author='Alexander Asp Bock',
author_email='albo.developer@gmail.com',
platforms='All',
description=('Bib(la)tex parsing and useful tools'),
license='MIT',
keywords='bibpy, bibtex, biblatex, parser',
url='https://github.com/MisanthropicBit/bibpy',
packages=['bibpy', 'bibpy.entry', 'bibpy.lexers', 'bibpy.parsers',
'bibpy.doi'],
long_description=open('README.md').read(),
scripts=['bin/bibgrep', 'bin/bibformat', 'bin/bibstats'],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Topic :: Utilities',
'Topic :: Software Development',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.5'
'Programming Language :: Python :: 3.6',
]
)
| """bibpy module setup script for distribution."""
from __future__ import with_statement
import os
import distutils.core
def get_version(filename):
with open(filename) as fh:
for line in fh:
if line.startswith('__version__'):
return line.split('=')[-1].strip()[1:-1]
distutils.core.setup(
name='bibpy',
version=get_version(os.path.join('bibpy', '__init__.py')),
author='Alexander Asp Bock',
author_email='albo.developer@gmail.com',
platforms='All',
description=('Bib(la)tex parsing and useful tools'),
license='MIT',
keywords='bibpy, bibtex, biblatex, parser',
url='https://github.com/MisanthropicBit/bibpy',
packages=['bibpy', 'bibpy.entry', 'bibpy.lexers', 'bibpy.doi'],
long_description=open('README.md').read(),
scripts=['bin/bibgrep', 'bin/bibformat', 'bin/bibstats'],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Topic :: Utilities',
'Topic :: Software Development',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.5'
'Programming Language :: Python :: 3.6',
]
)
| Remove 'bibpy.parsers' from package list | Remove 'bibpy.parsers' from package list
| Python | mit | MisanthropicBit/bibpy,MisanthropicBit/bibpy | """bibpy module setup script for distribution."""
from __future__ import with_statement
import os
import distutils.core
def get_version(filename):
with open(filename) as fh:
for line in fh:
if line.startswith('__version__'):
return line.split('=')[-1].strip()[1:-1]
distutils.core.setup(
name='bibpy',
version=get_version(os.path.join('bibpy', '__init__.py')),
author='Alexander Asp Bock',
author_email='albo.developer@gmail.com',
platforms='All',
description=('Bib(la)tex parsing and useful tools'),
license='MIT',
keywords='bibpy, bibtex, biblatex, parser',
url='https://github.com/MisanthropicBit/bibpy',
- packages=['bibpy', 'bibpy.entry', 'bibpy.lexers', 'bibpy.parsers',
+ packages=['bibpy', 'bibpy.entry', 'bibpy.lexers', 'bibpy.doi'],
- 'bibpy.doi'],
long_description=open('README.md').read(),
scripts=['bin/bibgrep', 'bin/bibformat', 'bin/bibstats'],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Topic :: Utilities',
'Topic :: Software Development',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.5'
'Programming Language :: Python :: 3.6',
]
)
| Remove 'bibpy.parsers' from package list | ## Code Before:
"""bibpy module setup script for distribution."""
from __future__ import with_statement
import os
import distutils.core
def get_version(filename):
with open(filename) as fh:
for line in fh:
if line.startswith('__version__'):
return line.split('=')[-1].strip()[1:-1]
distutils.core.setup(
name='bibpy',
version=get_version(os.path.join('bibpy', '__init__.py')),
author='Alexander Asp Bock',
author_email='albo.developer@gmail.com',
platforms='All',
description=('Bib(la)tex parsing and useful tools'),
license='MIT',
keywords='bibpy, bibtex, biblatex, parser',
url='https://github.com/MisanthropicBit/bibpy',
packages=['bibpy', 'bibpy.entry', 'bibpy.lexers', 'bibpy.parsers',
'bibpy.doi'],
long_description=open('README.md').read(),
scripts=['bin/bibgrep', 'bin/bibformat', 'bin/bibstats'],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Topic :: Utilities',
'Topic :: Software Development',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.5'
'Programming Language :: Python :: 3.6',
]
)
## Instruction:
Remove 'bibpy.parsers' from package list
## Code After:
"""bibpy module setup script for distribution."""
from __future__ import with_statement
import os
import distutils.core
def get_version(filename):
with open(filename) as fh:
for line in fh:
if line.startswith('__version__'):
return line.split('=')[-1].strip()[1:-1]
distutils.core.setup(
name='bibpy',
version=get_version(os.path.join('bibpy', '__init__.py')),
author='Alexander Asp Bock',
author_email='albo.developer@gmail.com',
platforms='All',
description=('Bib(la)tex parsing and useful tools'),
license='MIT',
keywords='bibpy, bibtex, biblatex, parser',
url='https://github.com/MisanthropicBit/bibpy',
packages=['bibpy', 'bibpy.entry', 'bibpy.lexers', 'bibpy.doi'],
long_description=open('README.md').read(),
scripts=['bin/bibgrep', 'bin/bibformat', 'bin/bibstats'],
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Topic :: Utilities',
'Topic :: Software Development',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.5'
'Programming Language :: Python :: 3.6',
]
)
|
73f7502f1fda11bc23469c6f3e8f79b0e375c928 | setup.py | setup.py |
from distutils.core import setup
setup(name='redis-dump-load',
version='0.2.0',
description='Dump and load redis databases',
author='Oleg Pudeyev',
author_email='oleg@bsdpower.com',
url='http://github.com/p/redis-dump-load',
py_modules=['redisdl'],
)
|
from distutils.core import setup
setup(name='redis-dump-load',
version='0.2.0',
description='Dump and load redis databases',
author='Oleg Pudeyev',
author_email='oleg@bsdpower.com',
url='http://github.com/p/redis-dump-load',
py_modules=['redisdl'],
data_files=['LICENSE', 'README.rst'],
)
| Add license and readme to the packages | Add license and readme to the packages
| Python | bsd-2-clause | hyunchel/redis-dump-load,p/redis-dump-load,p/redis-dump-load,hyunchel/redis-dump-load |
from distutils.core import setup
setup(name='redis-dump-load',
version='0.2.0',
description='Dump and load redis databases',
author='Oleg Pudeyev',
author_email='oleg@bsdpower.com',
url='http://github.com/p/redis-dump-load',
py_modules=['redisdl'],
+ data_files=['LICENSE', 'README.rst'],
)
| Add license and readme to the packages | ## Code Before:
from distutils.core import setup
setup(name='redis-dump-load',
version='0.2.0',
description='Dump and load redis databases',
author='Oleg Pudeyev',
author_email='oleg@bsdpower.com',
url='http://github.com/p/redis-dump-load',
py_modules=['redisdl'],
)
## Instruction:
Add license and readme to the packages
## Code After:
from distutils.core import setup
setup(name='redis-dump-load',
version='0.2.0',
description='Dump and load redis databases',
author='Oleg Pudeyev',
author_email='oleg@bsdpower.com',
url='http://github.com/p/redis-dump-load',
py_modules=['redisdl'],
data_files=['LICENSE', 'README.rst'],
)
|
c987ed375da13f53928157f14528bed0c148eeac | tasks.py | tasks.py | import asyncio
import threading
class Tasks:
loop = asyncio.new_event_loop()
@classmethod
def _run(cls):
try:
cls.loop.run_forever()
finally:
cls.loop.close()
@classmethod
def do(cls, func, *args, **kwargs):
cls.loop.call_soon(lambda: func(*args, **kwargs))
cls.loop._write_to_self()
@classmethod
def later(cls, func, *args, after=None, **kwargs):
cls.loop.call_later(after, lambda: func(*args, **kwargs))
cls.loop._write_to_self()
@classmethod
def periodic(cls, func, *args, interval=None, **kwargs):
@asyncio.coroutine
def f():
while True:
yield from asyncio.sleep(interval)
func(*args, **kwargs)
cls.loop.create_task(f())
cls.loop._write_to_self()
threading.Thread(name="tasks", target=Tasks._run, daemon=True).start()
| import asyncio
import threading
class Tasks:
loop = asyncio.new_event_loop()
@classmethod
def _run(cls):
asyncio.set_event_loop(cls.loop)
try:
cls.loop.run_forever()
finally:
cls.loop.close()
@classmethod
def do(cls, func, *args, **kwargs):
cls.loop.call_soon(lambda: func(*args, **kwargs))
cls.loop._write_to_self()
@classmethod
def later(cls, func, *args, after=None, **kwargs):
cls.loop.call_later(after, lambda: func(*args, **kwargs))
cls.loop._write_to_self()
@classmethod
def periodic(cls, func, *args, interval=None, **kwargs):
@asyncio.coroutine
def f():
while True:
yield from asyncio.sleep(interval)
func(*args, **kwargs)
cls.loop.create_task(f())
cls.loop._write_to_self()
threading.Thread(name="tasks", target=Tasks._run, daemon=True).start()
| Set implicit loop for Python <3.6 | Set implicit loop for Python <3.6 | Python | apache-2.0 | Charcoal-SE/SmokeDetector,Charcoal-SE/SmokeDetector | import asyncio
import threading
class Tasks:
loop = asyncio.new_event_loop()
@classmethod
def _run(cls):
+ asyncio.set_event_loop(cls.loop)
+
try:
cls.loop.run_forever()
finally:
cls.loop.close()
@classmethod
def do(cls, func, *args, **kwargs):
cls.loop.call_soon(lambda: func(*args, **kwargs))
cls.loop._write_to_self()
@classmethod
def later(cls, func, *args, after=None, **kwargs):
cls.loop.call_later(after, lambda: func(*args, **kwargs))
cls.loop._write_to_self()
@classmethod
def periodic(cls, func, *args, interval=None, **kwargs):
@asyncio.coroutine
def f():
while True:
yield from asyncio.sleep(interval)
func(*args, **kwargs)
cls.loop.create_task(f())
cls.loop._write_to_self()
threading.Thread(name="tasks", target=Tasks._run, daemon=True).start()
| Set implicit loop for Python <3.6 | ## Code Before:
import asyncio
import threading
class Tasks:
loop = asyncio.new_event_loop()
@classmethod
def _run(cls):
try:
cls.loop.run_forever()
finally:
cls.loop.close()
@classmethod
def do(cls, func, *args, **kwargs):
cls.loop.call_soon(lambda: func(*args, **kwargs))
cls.loop._write_to_self()
@classmethod
def later(cls, func, *args, after=None, **kwargs):
cls.loop.call_later(after, lambda: func(*args, **kwargs))
cls.loop._write_to_self()
@classmethod
def periodic(cls, func, *args, interval=None, **kwargs):
@asyncio.coroutine
def f():
while True:
yield from asyncio.sleep(interval)
func(*args, **kwargs)
cls.loop.create_task(f())
cls.loop._write_to_self()
threading.Thread(name="tasks", target=Tasks._run, daemon=True).start()
## Instruction:
Set implicit loop for Python <3.6
## Code After:
import asyncio
import threading
class Tasks:
loop = asyncio.new_event_loop()
@classmethod
def _run(cls):
asyncio.set_event_loop(cls.loop)
try:
cls.loop.run_forever()
finally:
cls.loop.close()
@classmethod
def do(cls, func, *args, **kwargs):
cls.loop.call_soon(lambda: func(*args, **kwargs))
cls.loop._write_to_self()
@classmethod
def later(cls, func, *args, after=None, **kwargs):
cls.loop.call_later(after, lambda: func(*args, **kwargs))
cls.loop._write_to_self()
@classmethod
def periodic(cls, func, *args, interval=None, **kwargs):
@asyncio.coroutine
def f():
while True:
yield from asyncio.sleep(interval)
func(*args, **kwargs)
cls.loop.create_task(f())
cls.loop._write_to_self()
threading.Thread(name="tasks", target=Tasks._run, daemon=True).start()
|
7dcd2c2aa1e2fd8f17e0b564f9b77375675ccd9a | metakernel/pexpect.py | metakernel/pexpect.py | from __future__ import absolute_import
from pexpect import spawn, which, EOF, TIMEOUT
| from __future__ import absolute_import
from pexpect import which as which_base, is_executable_file, EOF, TIMEOUT
import os
try:
from pexpect import spawn
import pty
except ImportError:
pty = None
def which(filename):
'''This takes a given filename; tries to find it in the environment path;
then checks if it is executable. This returns the full path to the filename
if found and executable. Otherwise this returns None.'''
# Special case where filename contains an explicit path.
if os.path.dirname(filename) != '' and is_executable_file(filename):
return filename
if 'PATH' not in os.environ or os.environ['PATH'] == '':
p = os.defpath
else:
p = os.environ['PATH']
pathlist = p.split(os.pathsep)
for path in pathlist:
ff = os.path.join(path, filename)
if pty:
if is_executable_file(ff):
return ff
else:
pathext = os.environ.get('Pathext', '.exe;.com;.bat;.cmd')
pathext = pathext.split(os.pathsep) + ['']
for ext in pathext:
if os.access(ff + ext, os.X_OK):
return ff + ext
return None
| Add handling of which on Windows | Add handling of which on Windows
| Python | bsd-3-clause | Calysto/metakernel | from __future__ import absolute_import
- from pexpect import spawn, which, EOF, TIMEOUT
+ from pexpect import which as which_base, is_executable_file, EOF, TIMEOUT
+ import os
+ try:
+ from pexpect import spawn
+ import pty
+ except ImportError:
+ pty = None
+
+
+ def which(filename):
+ '''This takes a given filename; tries to find it in the environment path;
+ then checks if it is executable. This returns the full path to the filename
+ if found and executable. Otherwise this returns None.'''
+
+ # Special case where filename contains an explicit path.
+ if os.path.dirname(filename) != '' and is_executable_file(filename):
+ return filename
+ if 'PATH' not in os.environ or os.environ['PATH'] == '':
+ p = os.defpath
+ else:
+ p = os.environ['PATH']
+ pathlist = p.split(os.pathsep)
+ for path in pathlist:
+ ff = os.path.join(path, filename)
+ if pty:
+ if is_executable_file(ff):
+ return ff
+ else:
+ pathext = os.environ.get('Pathext', '.exe;.com;.bat;.cmd')
+ pathext = pathext.split(os.pathsep) + ['']
+ for ext in pathext:
+ if os.access(ff + ext, os.X_OK):
+ return ff + ext
+ return None
+ | Add handling of which on Windows | ## Code Before:
from __future__ import absolute_import
from pexpect import spawn, which, EOF, TIMEOUT
## Instruction:
Add handling of which on Windows
## Code After:
from __future__ import absolute_import
from pexpect import which as which_base, is_executable_file, EOF, TIMEOUT
import os
try:
from pexpect import spawn
import pty
except ImportError:
pty = None
def which(filename):
'''This takes a given filename; tries to find it in the environment path;
then checks if it is executable. This returns the full path to the filename
if found and executable. Otherwise this returns None.'''
# Special case where filename contains an explicit path.
if os.path.dirname(filename) != '' and is_executable_file(filename):
return filename
if 'PATH' not in os.environ or os.environ['PATH'] == '':
p = os.defpath
else:
p = os.environ['PATH']
pathlist = p.split(os.pathsep)
for path in pathlist:
ff = os.path.join(path, filename)
if pty:
if is_executable_file(ff):
return ff
else:
pathext = os.environ.get('Pathext', '.exe;.com;.bat;.cmd')
pathext = pathext.split(os.pathsep) + ['']
for ext in pathext:
if os.access(ff + ext, os.X_OK):
return ff + ext
return None
|
6defa096b3dae109bf50ab32cdee7062c8b4327b | _python/config/settings/settings_pytest.py | _python/config/settings/settings_pytest.py |
from .settings_dev import *
# Don't use whitenoise for tests. Including whitenoise causes it to rescan static during each test, which greatly
# increases test time.
MIDDLEWARE.remove('whitenoise.middleware.WhiteNoiseMiddleware')
|
from .settings_dev import *
# Don't use whitenoise for tests. Including whitenoise causes it to rescan static during each test, which greatly
# increases test time.
MIDDLEWARE.remove('whitenoise.middleware.WhiteNoiseMiddleware')
CAPAPI_API_KEY = '12345'
| Add placeholder CAPAPI key for tests. | Add placeholder CAPAPI key for tests.
| Python | agpl-3.0 | harvard-lil/h2o,harvard-lil/h2o,harvard-lil/h2o,harvard-lil/h2o |
from .settings_dev import *
# Don't use whitenoise for tests. Including whitenoise causes it to rescan static during each test, which greatly
# increases test time.
MIDDLEWARE.remove('whitenoise.middleware.WhiteNoiseMiddleware')
+ CAPAPI_API_KEY = '12345'
| Add placeholder CAPAPI key for tests. | ## Code Before:
from .settings_dev import *
# Don't use whitenoise for tests. Including whitenoise causes it to rescan static during each test, which greatly
# increases test time.
MIDDLEWARE.remove('whitenoise.middleware.WhiteNoiseMiddleware')
## Instruction:
Add placeholder CAPAPI key for tests.
## Code After:
from .settings_dev import *
# Don't use whitenoise for tests. Including whitenoise causes it to rescan static during each test, which greatly
# increases test time.
MIDDLEWARE.remove('whitenoise.middleware.WhiteNoiseMiddleware')
CAPAPI_API_KEY = '12345'
|
d413747e996326e62fdf942426f170f66d5acb7c | osf_tests/test_preprint_summary.py | osf_tests/test_preprint_summary.py | import datetime
from osf_tests.factories import PreprintFactory, PreprintProviderFactory
from osf.models import PreprintService
from nose.tools import * # PEP8 asserts
import mock
import pytest
import pytz
import requests
from scripts.analytics.preprint_summary import PreprintSummary
@pytest.fixture()
def preprint_provider():
return PreprintProviderFactory(name='Test 1')
@pytest.fixture()
def preprint(preprint_provider):
return PreprintFactory._build(PreprintService, provider=preprint_provider)
pytestmark = pytest.mark.django_db
class TestPreprintCount:
def test_get_preprint_count(self, preprint_provider, preprint):
requests.post = mock.MagicMock()
resp = requests.Response()
resp._content = '{"hits" : {"total" : 1}}'
requests.post.return_value = resp
field = PreprintService._meta.get_field('date_created')
field.auto_now_add = False # We have to fudge the time because Keen doesn't allow same day queries.
date = datetime.datetime.utcnow() - datetime.timedelta(1)
preprint.date_created = date - datetime.timedelta(0.1)
preprint.save()
field.auto_now_add = True
results = PreprintSummary().get_events(date.date())
assert_equal(len(results), 1)
data = results[0]
assert_equal(data['provider']['name'], 'Test 1')
assert_equal(data['provider']['total'], 1)
| import datetime
from osf_tests.factories import PreprintFactory, PreprintProviderFactory
from osf.models import PreprintService
from nose.tools import * # PEP8 asserts
import mock
import pytest
import pytz
import requests
from scripts.analytics.preprint_summary import PreprintSummary
@pytest.fixture()
def preprint_provider():
return PreprintProviderFactory(name='Test 1')
@pytest.fixture()
def preprint(preprint_provider):
return PreprintFactory._build(PreprintService, provider=preprint_provider)
pytestmark = pytest.mark.django_db
class TestPreprintCount:
def test_get_preprint_count(self, preprint):
requests.post = mock.MagicMock()
resp = requests.Response()
resp._content = '{"hits" : {"total" : 1}}'
requests.post.return_value = resp
field = PreprintService._meta.get_field('date_created')
field.auto_now_add = False # We have to fudge the time because Keen doesn't allow same day queries.
date = datetime.datetime.utcnow() - datetime.timedelta(days=1, hours=1)
preprint.date_created = date - datetime.timedelta(hours=1)
preprint.save()
field.auto_now_add = True
results = PreprintSummary().get_events(date.date())
assert_equal(len(results), 1)
data = results[0]
assert_equal(data['provider']['name'], 'Test 1')
assert_equal(data['provider']['total'], 1)
| Make sure test dates are rounded properly by making they are over a day in the past. | Make sure test dates are rounded properly by making they are over a day in the past.
| Python | apache-2.0 | binoculars/osf.io,TomBaxter/osf.io,sloria/osf.io,adlius/osf.io,Johnetordoff/osf.io,mattclark/osf.io,adlius/osf.io,sloria/osf.io,HalcyonChimera/osf.io,adlius/osf.io,laurenrevere/osf.io,erinspace/osf.io,sloria/osf.io,binoculars/osf.io,mfraezz/osf.io,TomBaxter/osf.io,icereval/osf.io,Johnetordoff/osf.io,pattisdr/osf.io,crcresearch/osf.io,adlius/osf.io,felliott/osf.io,caseyrollins/osf.io,crcresearch/osf.io,aaxelb/osf.io,HalcyonChimera/osf.io,leb2dg/osf.io,chennan47/osf.io,cslzchen/osf.io,cslzchen/osf.io,Johnetordoff/osf.io,leb2dg/osf.io,erinspace/osf.io,leb2dg/osf.io,brianjgeiger/osf.io,caseyrollins/osf.io,CenterForOpenScience/osf.io,aaxelb/osf.io,cslzchen/osf.io,chennan47/osf.io,CenterForOpenScience/osf.io,HalcyonChimera/osf.io,mattclark/osf.io,felliott/osf.io,felliott/osf.io,laurenrevere/osf.io,mfraezz/osf.io,HalcyonChimera/osf.io,brianjgeiger/osf.io,mfraezz/osf.io,pattisdr/osf.io,pattisdr/osf.io,icereval/osf.io,laurenrevere/osf.io,baylee-d/osf.io,chennan47/osf.io,TomBaxter/osf.io,aaxelb/osf.io,binoculars/osf.io,crcresearch/osf.io,Johnetordoff/osf.io,aaxelb/osf.io,cslzchen/osf.io,felliott/osf.io,mfraezz/osf.io,saradbowman/osf.io,CenterForOpenScience/osf.io,mattclark/osf.io,baylee-d/osf.io,brianjgeiger/osf.io,erinspace/osf.io,brianjgeiger/osf.io,icereval/osf.io,caseyrollins/osf.io,CenterForOpenScience/osf.io,baylee-d/osf.io,leb2dg/osf.io,saradbowman/osf.io | import datetime
from osf_tests.factories import PreprintFactory, PreprintProviderFactory
from osf.models import PreprintService
from nose.tools import * # PEP8 asserts
import mock
import pytest
import pytz
import requests
from scripts.analytics.preprint_summary import PreprintSummary
@pytest.fixture()
def preprint_provider():
return PreprintProviderFactory(name='Test 1')
@pytest.fixture()
def preprint(preprint_provider):
return PreprintFactory._build(PreprintService, provider=preprint_provider)
pytestmark = pytest.mark.django_db
class TestPreprintCount:
- def test_get_preprint_count(self, preprint_provider, preprint):
+ def test_get_preprint_count(self, preprint):
requests.post = mock.MagicMock()
resp = requests.Response()
resp._content = '{"hits" : {"total" : 1}}'
requests.post.return_value = resp
field = PreprintService._meta.get_field('date_created')
field.auto_now_add = False # We have to fudge the time because Keen doesn't allow same day queries.
- date = datetime.datetime.utcnow() - datetime.timedelta(1)
+ date = datetime.datetime.utcnow() - datetime.timedelta(days=1, hours=1)
- preprint.date_created = date - datetime.timedelta(0.1)
+ preprint.date_created = date - datetime.timedelta(hours=1)
preprint.save()
field.auto_now_add = True
results = PreprintSummary().get_events(date.date())
assert_equal(len(results), 1)
data = results[0]
assert_equal(data['provider']['name'], 'Test 1')
assert_equal(data['provider']['total'], 1)
| Make sure test dates are rounded properly by making they are over a day in the past. | ## Code Before:
import datetime
from osf_tests.factories import PreprintFactory, PreprintProviderFactory
from osf.models import PreprintService
from nose.tools import * # PEP8 asserts
import mock
import pytest
import pytz
import requests
from scripts.analytics.preprint_summary import PreprintSummary
@pytest.fixture()
def preprint_provider():
return PreprintProviderFactory(name='Test 1')
@pytest.fixture()
def preprint(preprint_provider):
return PreprintFactory._build(PreprintService, provider=preprint_provider)
pytestmark = pytest.mark.django_db
class TestPreprintCount:
def test_get_preprint_count(self, preprint_provider, preprint):
requests.post = mock.MagicMock()
resp = requests.Response()
resp._content = '{"hits" : {"total" : 1}}'
requests.post.return_value = resp
field = PreprintService._meta.get_field('date_created')
field.auto_now_add = False # We have to fudge the time because Keen doesn't allow same day queries.
date = datetime.datetime.utcnow() - datetime.timedelta(1)
preprint.date_created = date - datetime.timedelta(0.1)
preprint.save()
field.auto_now_add = True
results = PreprintSummary().get_events(date.date())
assert_equal(len(results), 1)
data = results[0]
assert_equal(data['provider']['name'], 'Test 1')
assert_equal(data['provider']['total'], 1)
## Instruction:
Make sure test dates are rounded properly by making they are over a day in the past.
## Code After:
import datetime
from osf_tests.factories import PreprintFactory, PreprintProviderFactory
from osf.models import PreprintService
from nose.tools import * # PEP8 asserts
import mock
import pytest
import pytz
import requests
from scripts.analytics.preprint_summary import PreprintSummary
@pytest.fixture()
def preprint_provider():
return PreprintProviderFactory(name='Test 1')
@pytest.fixture()
def preprint(preprint_provider):
return PreprintFactory._build(PreprintService, provider=preprint_provider)
pytestmark = pytest.mark.django_db
class TestPreprintCount:
def test_get_preprint_count(self, preprint):
requests.post = mock.MagicMock()
resp = requests.Response()
resp._content = '{"hits" : {"total" : 1}}'
requests.post.return_value = resp
field = PreprintService._meta.get_field('date_created')
field.auto_now_add = False # We have to fudge the time because Keen doesn't allow same day queries.
date = datetime.datetime.utcnow() - datetime.timedelta(days=1, hours=1)
preprint.date_created = date - datetime.timedelta(hours=1)
preprint.save()
field.auto_now_add = True
results = PreprintSummary().get_events(date.date())
assert_equal(len(results), 1)
data = results[0]
assert_equal(data['provider']['name'], 'Test 1')
assert_equal(data['provider']['total'], 1)
|
7d6ad23cd8435eac9b48a4ea63bb9a2e83239c4a | scent.py | scent.py |
import os
import termstyle
from sniffer.api import file_validator, runnable
from tmuxp.testsuite import main
# you can customize the pass/fail colors like this
pass_fg_color = termstyle.green
pass_bg_color = termstyle.bg_default
fail_fg_color = termstyle.red
fail_bg_color = termstyle.bg_default
# All lists in this variable will be under surveillance for changes.
watch_paths = ['tmuxp/']
@file_validator
def py_files(filename):
return filename.endswith('.py') and not os.path.basename(filename).startswith('.') and filename != ".tmuxp"
@runnable
def execute_nose(*args):
try:
return main()
except SystemExit as x:
if x.message:
print "Found error {0}: {1}".format(x.code, x.message)
return not x.code
else:
return 1
| from __future__ import unicode_literals
import os
import termstyle
from sniffer.api import file_validator, runnable
from tmuxp.testsuite import main
# you can customize the pass/fail colors like this
pass_fg_color = termstyle.green
pass_bg_color = termstyle.bg_default
fail_fg_color = termstyle.red
fail_bg_color = termstyle.bg_default
# All lists in this variable will be under surveillance for changes.
watch_paths = ['tmuxp/']
@file_validator
def py_files(filename):
return filename.endswith('.py') and not os.path.basename(filename).startswith('.') and filename != ".tmuxp"
@runnable
def execute_nose(*args):
try:
return main()
except SystemExit as x:
if x.message:
print("Found error {0}: {1}".format(x.code, x.message))
return not x.code
else:
return 1
| Fix sniffer support for python 3.x | Fix sniffer support for python 3.x
| Python | bsd-3-clause | thomasballinger/tmuxp,thomasballinger/tmuxp,mexicarne/tmuxp,tony/tmuxp,mexicarne/tmuxp | + from __future__ import unicode_literals
import os
import termstyle
from sniffer.api import file_validator, runnable
from tmuxp.testsuite import main
# you can customize the pass/fail colors like this
pass_fg_color = termstyle.green
pass_bg_color = termstyle.bg_default
fail_fg_color = termstyle.red
fail_bg_color = termstyle.bg_default
# All lists in this variable will be under surveillance for changes.
watch_paths = ['tmuxp/']
@file_validator
def py_files(filename):
return filename.endswith('.py') and not os.path.basename(filename).startswith('.') and filename != ".tmuxp"
@runnable
def execute_nose(*args):
try:
return main()
except SystemExit as x:
if x.message:
- print "Found error {0}: {1}".format(x.code, x.message)
+ print("Found error {0}: {1}".format(x.code, x.message))
return not x.code
else:
return 1
- | Fix sniffer support for python 3.x | ## Code Before:
import os
import termstyle
from sniffer.api import file_validator, runnable
from tmuxp.testsuite import main
# you can customize the pass/fail colors like this
pass_fg_color = termstyle.green
pass_bg_color = termstyle.bg_default
fail_fg_color = termstyle.red
fail_bg_color = termstyle.bg_default
# All lists in this variable will be under surveillance for changes.
watch_paths = ['tmuxp/']
@file_validator
def py_files(filename):
return filename.endswith('.py') and not os.path.basename(filename).startswith('.') and filename != ".tmuxp"
@runnable
def execute_nose(*args):
try:
return main()
except SystemExit as x:
if x.message:
print "Found error {0}: {1}".format(x.code, x.message)
return not x.code
else:
return 1
## Instruction:
Fix sniffer support for python 3.x
## Code After:
from __future__ import unicode_literals
import os
import termstyle
from sniffer.api import file_validator, runnable
from tmuxp.testsuite import main
# you can customize the pass/fail colors like this
pass_fg_color = termstyle.green
pass_bg_color = termstyle.bg_default
fail_fg_color = termstyle.red
fail_bg_color = termstyle.bg_default
# All lists in this variable will be under surveillance for changes.
watch_paths = ['tmuxp/']
@file_validator
def py_files(filename):
return filename.endswith('.py') and not os.path.basename(filename).startswith('.') and filename != ".tmuxp"
@runnable
def execute_nose(*args):
try:
return main()
except SystemExit as x:
if x.message:
print("Found error {0}: {1}".format(x.code, x.message))
return not x.code
else:
return 1
|
ec96c3173a770949c13e560b16272bc265a80da4 | setup.py | setup.py |
from distutils.core import setup
setup(name='mass_api_client',
version=0.1,
install_required=['requests==2.13.0', 'marshmallow==2.12.2'])
|
from distutils.core import setup
setup(name='mass_api_client',
version=0.1,
install_requires=['requests==2.13.0', 'marshmallow==2.12.2'],
packages=['mass_api_client', ],
)
| Add mass_api_client as Package; fix typo | Add mass_api_client as Package; fix typo
| Python | mit | mass-project/mass_api_client,mass-project/mass_api_client |
from distutils.core import setup
setup(name='mass_api_client',
version=0.1,
- install_required=['requests==2.13.0', 'marshmallow==2.12.2'])
+ install_requires=['requests==2.13.0', 'marshmallow==2.12.2'],
+ packages=['mass_api_client', ],
+ )
| Add mass_api_client as Package; fix typo | ## Code Before:
from distutils.core import setup
setup(name='mass_api_client',
version=0.1,
install_required=['requests==2.13.0', 'marshmallow==2.12.2'])
## Instruction:
Add mass_api_client as Package; fix typo
## Code After:
from distutils.core import setup
setup(name='mass_api_client',
version=0.1,
install_requires=['requests==2.13.0', 'marshmallow==2.12.2'],
packages=['mass_api_client', ],
)
|
596613c964311104098e64eeb349216bc7cd0023 | saleor/demo/views.py | saleor/demo/views.py | from django.conf import settings
from django.shortcuts import render
from ..graphql.views import API_PATH, GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
ctx = {
"query": EXAMPLE_QUERY,
"api_url": request.build_absolute_uri(str(API_PATH)),
}
return render(request, "graphql/playground.html", ctx)
| from django.conf import settings
from django.shortcuts import render
from ..graphql.views import GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
pwa_origin = settings.PWA_ORIGINS[0]
ctx = {
"query": EXAMPLE_QUERY,
"api_url": f"https://{pwa_origin}/graphql/",
}
return render(request, "graphql/playground.html", ctx)
| Fix playground CSP for demo if deployed under proxied domain | Fix playground CSP for demo if deployed under proxied domain
| Python | bsd-3-clause | mociepka/saleor,mociepka/saleor,mociepka/saleor | from django.conf import settings
from django.shortcuts import render
- from ..graphql.views import API_PATH, GraphQLView
+ from ..graphql.views import GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
+ pwa_origin = settings.PWA_ORIGINS[0]
ctx = {
"query": EXAMPLE_QUERY,
- "api_url": request.build_absolute_uri(str(API_PATH)),
+ "api_url": f"https://{pwa_origin}/graphql/",
}
return render(request, "graphql/playground.html", ctx)
| Fix playground CSP for demo if deployed under proxied domain | ## Code Before:
from django.conf import settings
from django.shortcuts import render
from ..graphql.views import API_PATH, GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
ctx = {
"query": EXAMPLE_QUERY,
"api_url": request.build_absolute_uri(str(API_PATH)),
}
return render(request, "graphql/playground.html", ctx)
## Instruction:
Fix playground CSP for demo if deployed under proxied domain
## Code After:
from django.conf import settings
from django.shortcuts import render
from ..graphql.views import GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
pwa_origin = settings.PWA_ORIGINS[0]
ctx = {
"query": EXAMPLE_QUERY,
"api_url": f"https://{pwa_origin}/graphql/",
}
return render(request, "graphql/playground.html", ctx)
|
53594f372a45e425076e5dbf36f399503df1972c | salt/output/yaml_out.py | salt/output/yaml_out.py | '''
YAML Outputter
'''
# Third Party libs
import yaml
def __virtual__():
return 'yaml'
def output(data):
'''
Print out YAML
'''
return yaml.dump(data)
| '''
Output data in YAML, this outputter defaults to printing in YAML block mode
for better readability.
'''
# Third Party libs
import yaml
def __virtual__():
return 'yaml'
def output(data):
'''
Print out YAML using the block mode
'''
return yaml.dump(data, default_flow_style=False)
| Change the YAML outputter to use block mode and add some docs | Change the YAML outputter to use block mode and add some docs
| Python | apache-2.0 | saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt | '''
- YAML Outputter
+ Output data in YAML, this outputter defaults to printing in YAML block mode
+ for better readability.
'''
# Third Party libs
import yaml
def __virtual__():
return 'yaml'
def output(data):
'''
- Print out YAML
+ Print out YAML using the block mode
'''
- return yaml.dump(data)
+ return yaml.dump(data, default_flow_style=False)
| Change the YAML outputter to use block mode and add some docs | ## Code Before:
'''
YAML Outputter
'''
# Third Party libs
import yaml
def __virtual__():
return 'yaml'
def output(data):
'''
Print out YAML
'''
return yaml.dump(data)
## Instruction:
Change the YAML outputter to use block mode and add some docs
## Code After:
'''
Output data in YAML, this outputter defaults to printing in YAML block mode
for better readability.
'''
# Third Party libs
import yaml
def __virtual__():
return 'yaml'
def output(data):
'''
Print out YAML using the block mode
'''
return yaml.dump(data, default_flow_style=False)
|
eb5bcf3130f5fdc0d6be68e7e81555def46a53af | setup.py | setup.py | from distutils.core import setup
setup(
name = 'mstranslator',
packages = ['mstranslator'],
version = '0.0.1',
description = 'Python wrapper to consume Microsoft translator API',
author = 'Ayush Goel',
author_email = 'ayushgoel111@gmail.com',
url = 'https://github.com/ayushgoel/mstranslator',
download_url = 'https://github.com/peterldowns/mypackage/tarball/0.1',
keywords = ['microsoft', 'translator', 'language'],
requires = ['requests']
)
| from distutils.core import setup
setup(
name = 'mstranslator-2016',
packages = ['mstranslator'],
version = '0.0.1',
description = 'Python wrapper to consume Microsoft translator API',
author = 'Ayush Goel',
author_email = 'ayushgoel111@gmail.com',
url = 'https://github.com/ayushgoel/mstranslator',
download_url = 'https://github.com/ayushgoel/mstranslator/archive/0.0.1.tar.gz',
keywords = ['microsoft', 'translator', 'language'],
requires = ['requests']
)
| Update nemae and download URL | Update nemae and download URL
| Python | mit | ayushgoel/mstranslator | from distutils.core import setup
setup(
- name = 'mstranslator',
+ name = 'mstranslator-2016',
packages = ['mstranslator'],
version = '0.0.1',
description = 'Python wrapper to consume Microsoft translator API',
author = 'Ayush Goel',
author_email = 'ayushgoel111@gmail.com',
url = 'https://github.com/ayushgoel/mstranslator',
- download_url = 'https://github.com/peterldowns/mypackage/tarball/0.1',
+ download_url = 'https://github.com/ayushgoel/mstranslator/archive/0.0.1.tar.gz',
keywords = ['microsoft', 'translator', 'language'],
requires = ['requests']
)
| Update nemae and download URL | ## Code Before:
from distutils.core import setup
setup(
name = 'mstranslator',
packages = ['mstranslator'],
version = '0.0.1',
description = 'Python wrapper to consume Microsoft translator API',
author = 'Ayush Goel',
author_email = 'ayushgoel111@gmail.com',
url = 'https://github.com/ayushgoel/mstranslator',
download_url = 'https://github.com/peterldowns/mypackage/tarball/0.1',
keywords = ['microsoft', 'translator', 'language'],
requires = ['requests']
)
## Instruction:
Update nemae and download URL
## Code After:
from distutils.core import setup
setup(
name = 'mstranslator-2016',
packages = ['mstranslator'],
version = '0.0.1',
description = 'Python wrapper to consume Microsoft translator API',
author = 'Ayush Goel',
author_email = 'ayushgoel111@gmail.com',
url = 'https://github.com/ayushgoel/mstranslator',
download_url = 'https://github.com/ayushgoel/mstranslator/archive/0.0.1.tar.gz',
keywords = ['microsoft', 'translator', 'language'],
requires = ['requests']
)
|
ccbe4a1c48765fdd9e785392dff949bcc49192a2 | setup.py | setup.py | from distutils.core import setup
setup(
name='Zinc',
version='0.1.7',
author='John Wang',
author_email='john@zinc.io',
packages=['zinc'],
package_dir={'zinc': ''},
package_data={'zinc': ['examples/*.py', 'examples/*.json', 'README', 'zinc/*']},
include_package_data=True,
url='https://github.com/wangjohn/zinc_cli',
license='LICENSE.txt',
description='Wrapper for Zinc ecommerce API (zinc.io)',
install_requires=[
"requests >= 1.1.0"
],
)
| from distutils.core import setup
setup(
name='Zinc',
version='0.1.8',
author='John Wang',
author_email='john@zinc.io',
packages=['zinc'],
package_dir={'zinc': ''},
package_data={'zinc': ['examples/*.py', 'examples/*.json', 'zinc/*']},
include_package_data=True,
url='https://github.com/wangjohn/zinc_cli',
license='LICENSE.txt',
description='Wrapper for Zinc ecommerce API (zinc.io)',
install_requires=[
"requests >= 1.1.0"
],
)
| Remove readme from package data. | Remove readme from package data.
| Python | mit | wangjohn/zinc_cli | from distutils.core import setup
setup(
name='Zinc',
- version='0.1.7',
+ version='0.1.8',
author='John Wang',
author_email='john@zinc.io',
packages=['zinc'],
package_dir={'zinc': ''},
- package_data={'zinc': ['examples/*.py', 'examples/*.json', 'README', 'zinc/*']},
+ package_data={'zinc': ['examples/*.py', 'examples/*.json', 'zinc/*']},
include_package_data=True,
url='https://github.com/wangjohn/zinc_cli',
license='LICENSE.txt',
description='Wrapper for Zinc ecommerce API (zinc.io)',
install_requires=[
"requests >= 1.1.0"
],
)
| Remove readme from package data. | ## Code Before:
from distutils.core import setup
setup(
name='Zinc',
version='0.1.7',
author='John Wang',
author_email='john@zinc.io',
packages=['zinc'],
package_dir={'zinc': ''},
package_data={'zinc': ['examples/*.py', 'examples/*.json', 'README', 'zinc/*']},
include_package_data=True,
url='https://github.com/wangjohn/zinc_cli',
license='LICENSE.txt',
description='Wrapper for Zinc ecommerce API (zinc.io)',
install_requires=[
"requests >= 1.1.0"
],
)
## Instruction:
Remove readme from package data.
## Code After:
from distutils.core import setup
setup(
name='Zinc',
version='0.1.8',
author='John Wang',
author_email='john@zinc.io',
packages=['zinc'],
package_dir={'zinc': ''},
package_data={'zinc': ['examples/*.py', 'examples/*.json', 'zinc/*']},
include_package_data=True,
url='https://github.com/wangjohn/zinc_cli',
license='LICENSE.txt',
description='Wrapper for Zinc ecommerce API (zinc.io)',
install_requires=[
"requests >= 1.1.0"
],
)
|
7e88739d91cd7db35ffb36804ae59d1878eb2da3 | setup.py | setup.py | import os
from distutils.core import setup
requirements = map(str.strip, open('requirements.txt').readlines())
setup(
name='py_eventsocket',
version='0.1.4',
author="Aaron Westendorf",
author_email="aaron@agoragames.com",
packages = ['eventsocket'],
url='https://github.com/agoragames/py-eventsocket',
license='LICENSE.txt',
description='Easy to use TCP socket based on libevent',
install_requires = requirements,
long_description=open('README.rst').read(),
keywords=['socket', 'event'],
classifiers=[
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
"Intended Audience :: Developers",
"Operating System :: POSIX",
"Topic :: Communications",
"Topic :: Software Development :: Libraries :: Python Modules",
'Programming Language :: Python',
'Topic :: Software Development :: Libraries'
]
)
| import os
from distutils.core import setup
requirements = map(str.strip, open('requirements.txt').readlines())
setup(
name='py_eventsocket',
version='0.1.4',
author="Aaron Westendorf",
author_email="aaron@agoragames.com",
url='https://github.com/agoragames/py-eventsocket',
license='LICENSE.txt',
py_modules = ['eventsocket'],
description='Easy to use TCP socket based on libevent',
install_requires = requirements,
long_description=open('README.rst').read(),
keywords=['socket', 'event'],
classifiers=[
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
"Intended Audience :: Developers",
"Operating System :: POSIX",
"Topic :: Communications",
"Topic :: Software Development :: Libraries :: Python Modules",
'Programming Language :: Python',
'Topic :: Software Development :: Libraries'
]
)
| Use py_modules and not packages | Use py_modules and not packages
| Python | bsd-3-clause | agoragames/py-eventsocket | import os
from distutils.core import setup
requirements = map(str.strip, open('requirements.txt').readlines())
setup(
name='py_eventsocket',
version='0.1.4',
author="Aaron Westendorf",
author_email="aaron@agoragames.com",
- packages = ['eventsocket'],
url='https://github.com/agoragames/py-eventsocket',
license='LICENSE.txt',
+ py_modules = ['eventsocket'],
description='Easy to use TCP socket based on libevent',
install_requires = requirements,
long_description=open('README.rst').read(),
keywords=['socket', 'event'],
classifiers=[
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
"Intended Audience :: Developers",
"Operating System :: POSIX",
"Topic :: Communications",
"Topic :: Software Development :: Libraries :: Python Modules",
'Programming Language :: Python',
'Topic :: Software Development :: Libraries'
]
)
| Use py_modules and not packages | ## Code Before:
import os
from distutils.core import setup
requirements = map(str.strip, open('requirements.txt').readlines())
setup(
name='py_eventsocket',
version='0.1.4',
author="Aaron Westendorf",
author_email="aaron@agoragames.com",
packages = ['eventsocket'],
url='https://github.com/agoragames/py-eventsocket',
license='LICENSE.txt',
description='Easy to use TCP socket based on libevent',
install_requires = requirements,
long_description=open('README.rst').read(),
keywords=['socket', 'event'],
classifiers=[
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
"Intended Audience :: Developers",
"Operating System :: POSIX",
"Topic :: Communications",
"Topic :: Software Development :: Libraries :: Python Modules",
'Programming Language :: Python',
'Topic :: Software Development :: Libraries'
]
)
## Instruction:
Use py_modules and not packages
## Code After:
import os
from distutils.core import setup
requirements = map(str.strip, open('requirements.txt').readlines())
setup(
name='py_eventsocket',
version='0.1.4',
author="Aaron Westendorf",
author_email="aaron@agoragames.com",
url='https://github.com/agoragames/py-eventsocket',
license='LICENSE.txt',
py_modules = ['eventsocket'],
description='Easy to use TCP socket based on libevent',
install_requires = requirements,
long_description=open('README.rst').read(),
keywords=['socket', 'event'],
classifiers=[
'Development Status :: 4 - Beta',
'License :: OSI Approved :: BSD License',
"Intended Audience :: Developers",
"Operating System :: POSIX",
"Topic :: Communications",
"Topic :: Software Development :: Libraries :: Python Modules",
'Programming Language :: Python',
'Topic :: Software Development :: Libraries'
]
)
|
c0b9c9712e464f304bee7c63bfd6b197a1c5fb0f | cmsplugin_bootstrap_carousel/cms_plugins.py | cmsplugin_bootstrap_carousel/cms_plugins.py | import re
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from cmsplugin_bootstrap_carousel.models import *
from django.utils.translation import ugettext as _
from django.contrib import admin
from django.forms import ModelForm, ValidationError
class CarouselForm(ModelForm):
class Meta:
model = Carousel
def clean_domid(self):
data = self.cleaned_data['domid']
if not re.match(r'^[a-zA-Z_]\w*$', data):
raise ValidationError(_("The name must be a single word beginning with a letter"))
return data
class CarouselItemInline(admin.StackedInline):
model = CarouselItem
class CarouselPlugin(CMSPluginBase):
model = Carousel
form = CarouselForm
name = _("Carousel")
render_template = "cmsplugin_bootstrap_carousel/carousel.html"
inlines = [
CarouselItemInline,
]
def render(self, context, instance, placeholder):
context.update({'instance' : instance})
return context
plugin_pool.register_plugin(CarouselPlugin)
| import re
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from cmsplugin_bootstrap_carousel.models import *
from django.utils.translation import ugettext as _
from django.contrib import admin
from django.forms import ModelForm, ValidationError
class CarouselForm(ModelForm):
class Meta:
model = Carousel
def clean_domid(self):
data = self.cleaned_data['domid']
if not re.match(r'^[a-zA-Z_]\w*$', data):
raise ValidationError(_("The name must be a single word beginning with a letter"))
return data
class CarouselItemInline(admin.StackedInline):
model = CarouselItem
extra = 0
class CarouselPlugin(CMSPluginBase):
model = Carousel
form = CarouselForm
name = _("Carousel")
render_template = "cmsplugin_bootstrap_carousel/carousel.html"
inlines = [
CarouselItemInline,
]
def render(self, context, instance, placeholder):
context.update({'instance' : instance})
return context
plugin_pool.register_plugin(CarouselPlugin)
| Change extra from 3 to 0. | Change extra from 3 to 0.
| Python | bsd-3-clause | 360youlun/cmsplugin-bootstrap-carousel,360youlun/cmsplugin-bootstrap-carousel | import re
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from cmsplugin_bootstrap_carousel.models import *
from django.utils.translation import ugettext as _
from django.contrib import admin
from django.forms import ModelForm, ValidationError
class CarouselForm(ModelForm):
class Meta:
model = Carousel
def clean_domid(self):
data = self.cleaned_data['domid']
if not re.match(r'^[a-zA-Z_]\w*$', data):
raise ValidationError(_("The name must be a single word beginning with a letter"))
return data
class CarouselItemInline(admin.StackedInline):
model = CarouselItem
+ extra = 0
class CarouselPlugin(CMSPluginBase):
model = Carousel
form = CarouselForm
name = _("Carousel")
render_template = "cmsplugin_bootstrap_carousel/carousel.html"
inlines = [
CarouselItemInline,
]
def render(self, context, instance, placeholder):
context.update({'instance' : instance})
return context
plugin_pool.register_plugin(CarouselPlugin)
| Change extra from 3 to 0. | ## Code Before:
import re
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from cmsplugin_bootstrap_carousel.models import *
from django.utils.translation import ugettext as _
from django.contrib import admin
from django.forms import ModelForm, ValidationError
class CarouselForm(ModelForm):
class Meta:
model = Carousel
def clean_domid(self):
data = self.cleaned_data['domid']
if not re.match(r'^[a-zA-Z_]\w*$', data):
raise ValidationError(_("The name must be a single word beginning with a letter"))
return data
class CarouselItemInline(admin.StackedInline):
model = CarouselItem
class CarouselPlugin(CMSPluginBase):
model = Carousel
form = CarouselForm
name = _("Carousel")
render_template = "cmsplugin_bootstrap_carousel/carousel.html"
inlines = [
CarouselItemInline,
]
def render(self, context, instance, placeholder):
context.update({'instance' : instance})
return context
plugin_pool.register_plugin(CarouselPlugin)
## Instruction:
Change extra from 3 to 0.
## Code After:
import re
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from cmsplugin_bootstrap_carousel.models import *
from django.utils.translation import ugettext as _
from django.contrib import admin
from django.forms import ModelForm, ValidationError
class CarouselForm(ModelForm):
class Meta:
model = Carousel
def clean_domid(self):
data = self.cleaned_data['domid']
if not re.match(r'^[a-zA-Z_]\w*$', data):
raise ValidationError(_("The name must be a single word beginning with a letter"))
return data
class CarouselItemInline(admin.StackedInline):
model = CarouselItem
extra = 0
class CarouselPlugin(CMSPluginBase):
model = Carousel
form = CarouselForm
name = _("Carousel")
render_template = "cmsplugin_bootstrap_carousel/carousel.html"
inlines = [
CarouselItemInline,
]
def render(self, context, instance, placeholder):
context.update({'instance' : instance})
return context
plugin_pool.register_plugin(CarouselPlugin)
|
554e79ada3f351ecb6287b08d0f7d1c4e5a5b5f6 | setup.py | setup.py |
import sys
from distutils.core import setup
setup_args = {}
setup_args.update(dict(
name='param',
version='0.05',
description='Declarative Python programming using Parameters.',
long_description=open('README.txt').read(),
author= "IOAM",
author_email= "developers@topographica.org",
maintainer= "IOAM",
maintainer_email= "developers@topographica.org",
platforms=['Windows', 'Mac OS X', 'Linux'],
license='BSD',
url='http://ioam.github.com/param/',
packages = ["param"],
classifiers = [
"License :: OSI Approved :: BSD License",
# (until packaging tested)
"Development Status :: 4 - Beta",
"Programming Language :: Python :: 2.5",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Operating System :: OS Independent",
"Intended Audience :: Science/Research",
"Intended Audience :: Developers",
"Natural Language :: English",
"Topic :: Scientific/Engineering",
"Topic :: Software Development :: Libraries"]
))
if __name__=="__main__":
setup(**setup_args)
|
import sys
from distutils.core import setup
setup_args = {}
setup_args.update(dict(
name='param',
version='1.0',
description='Declarative Python programming using Parameters.',
long_description=open('README.txt').read(),
author= "IOAM",
author_email= "developers@topographica.org",
maintainer= "IOAM",
maintainer_email= "developers@topographica.org",
platforms=['Windows', 'Mac OS X', 'Linux'],
license='BSD',
url='http://ioam.github.com/param/',
packages = ["param"],
classifiers = [
"License :: OSI Approved :: BSD License",
"Development Status :: 5 - Production/Stable",
"Programming Language :: Python :: 2.5",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Operating System :: OS Independent",
"Intended Audience :: Science/Research",
"Intended Audience :: Developers",
"Natural Language :: English",
"Topic :: Scientific/Engineering",
"Topic :: Software Development :: Libraries"]
))
if __name__=="__main__":
setup(**setup_args)
| Update version number to 1.0. | Update version number to 1.0.
| Python | bsd-3-clause | ceball/param,ioam/param |
import sys
from distutils.core import setup
setup_args = {}
setup_args.update(dict(
name='param',
- version='0.05',
+ version='1.0',
description='Declarative Python programming using Parameters.',
long_description=open('README.txt').read(),
author= "IOAM",
author_email= "developers@topographica.org",
maintainer= "IOAM",
maintainer_email= "developers@topographica.org",
platforms=['Windows', 'Mac OS X', 'Linux'],
license='BSD',
url='http://ioam.github.com/param/',
packages = ["param"],
classifiers = [
"License :: OSI Approved :: BSD License",
- # (until packaging tested)
- "Development Status :: 4 - Beta",
+ "Development Status :: 5 - Production/Stable",
"Programming Language :: Python :: 2.5",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Operating System :: OS Independent",
"Intended Audience :: Science/Research",
"Intended Audience :: Developers",
"Natural Language :: English",
"Topic :: Scientific/Engineering",
"Topic :: Software Development :: Libraries"]
))
if __name__=="__main__":
setup(**setup_args)
| Update version number to 1.0. | ## Code Before:
import sys
from distutils.core import setup
setup_args = {}
setup_args.update(dict(
name='param',
version='0.05',
description='Declarative Python programming using Parameters.',
long_description=open('README.txt').read(),
author= "IOAM",
author_email= "developers@topographica.org",
maintainer= "IOAM",
maintainer_email= "developers@topographica.org",
platforms=['Windows', 'Mac OS X', 'Linux'],
license='BSD',
url='http://ioam.github.com/param/',
packages = ["param"],
classifiers = [
"License :: OSI Approved :: BSD License",
# (until packaging tested)
"Development Status :: 4 - Beta",
"Programming Language :: Python :: 2.5",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Operating System :: OS Independent",
"Intended Audience :: Science/Research",
"Intended Audience :: Developers",
"Natural Language :: English",
"Topic :: Scientific/Engineering",
"Topic :: Software Development :: Libraries"]
))
if __name__=="__main__":
setup(**setup_args)
## Instruction:
Update version number to 1.0.
## Code After:
import sys
from distutils.core import setup
setup_args = {}
setup_args.update(dict(
name='param',
version='1.0',
description='Declarative Python programming using Parameters.',
long_description=open('README.txt').read(),
author= "IOAM",
author_email= "developers@topographica.org",
maintainer= "IOAM",
maintainer_email= "developers@topographica.org",
platforms=['Windows', 'Mac OS X', 'Linux'],
license='BSD',
url='http://ioam.github.com/param/',
packages = ["param"],
classifiers = [
"License :: OSI Approved :: BSD License",
"Development Status :: 5 - Production/Stable",
"Programming Language :: Python :: 2.5",
"Programming Language :: Python :: 2.6",
"Programming Language :: Python :: 2.7",
"Operating System :: OS Independent",
"Intended Audience :: Science/Research",
"Intended Audience :: Developers",
"Natural Language :: English",
"Topic :: Scientific/Engineering",
"Topic :: Software Development :: Libraries"]
))
if __name__=="__main__":
setup(**setup_args)
|
81b6a138c476084f9ddd6063f31d3efd0ba6e2cf | start.py | start.py | import argparse
import logging
import os
import sys
from twisted.internet import reactor
from desertbot.config import Config, ConfigError
from desertbot.factory import DesertBotFactory
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='An IRC bot written in Python.')
parser.add_argument('-c', '--config',
help='the config file to read from',
type=str, required=True)
cmdArgs = parser.parse_args()
os.chdir(os.path.dirname(os.path.abspath(__file__)))
# Set up logging for stdout on the root 'desertbot' logger
# Modules can then just add more handlers to the root logger to capture all logs to files in various ways
rootLogger = logging.getLogger('desertbot')
rootLogger.setLevel(logging.INFO) # TODO change this from config value once it's loaded
logFormatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s', '%H:%M:%S')
streamHandler = logging.StreamHandler(stream=sys.stdout)
streamHandler.setFormatter(logFormatter)
rootLogger.addHandler(streamHandler)
config = Config(cmdArgs.config)
try:
config.loadConfig()
except ConfigError:
rootLogger.exception("Failed to load configuration file {}".format(cmdArgs.config))
else:
factory = DesertBotFactory(config)
reactor.run()
| import argparse
import logging
import os
import sys
from twisted.internet import reactor
from desertbot.config import Config, ConfigError
from desertbot.factory import DesertBotFactory
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='An IRC bot written in Python.')
parser.add_argument('-c', '--config',
help='the config file to read from',
type=str, required=True)
parser.add_argument('-l', '--loglevel',
help='the logging level (default INFO)',
type=str, default='INFO')
cmdArgs = parser.parse_args()
os.chdir(os.path.dirname(os.path.abspath(__file__)))
# Set up logging for stdout on the root 'desertbot' logger
# Modules can then just add more handlers to the root logger to capture all logs to files in various ways
rootLogger = logging.getLogger('desertbot')
numericLevel = getattr(logging, cmdArgs.loglevel.upper(), None)
if isinstance(numericLevel, int):
rootLogger.setLevel(numericLevel)
else:
raise ValueError('Invalid log level {}'.format(cmdArgs.loglevel))
logFormatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s', '%H:%M:%S')
streamHandler = logging.StreamHandler(stream=sys.stdout)
streamHandler.setFormatter(logFormatter)
rootLogger.addHandler(streamHandler)
config = Config(cmdArgs.config)
try:
config.loadConfig()
except ConfigError:
rootLogger.exception("Failed to load configuration file {}".format(cmdArgs.config))
else:
factory = DesertBotFactory(config)
reactor.run()
| Make the logging level configurable | Make the logging level configurable
| Python | mit | DesertBot/DesertBot | import argparse
import logging
import os
import sys
from twisted.internet import reactor
from desertbot.config import Config, ConfigError
from desertbot.factory import DesertBotFactory
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='An IRC bot written in Python.')
parser.add_argument('-c', '--config',
help='the config file to read from',
type=str, required=True)
+ parser.add_argument('-l', '--loglevel',
+ help='the logging level (default INFO)',
+ type=str, default='INFO')
cmdArgs = parser.parse_args()
os.chdir(os.path.dirname(os.path.abspath(__file__)))
# Set up logging for stdout on the root 'desertbot' logger
# Modules can then just add more handlers to the root logger to capture all logs to files in various ways
rootLogger = logging.getLogger('desertbot')
- rootLogger.setLevel(logging.INFO) # TODO change this from config value once it's loaded
+ numericLevel = getattr(logging, cmdArgs.loglevel.upper(), None)
+ if isinstance(numericLevel, int):
+ rootLogger.setLevel(numericLevel)
+ else:
+ raise ValueError('Invalid log level {}'.format(cmdArgs.loglevel))
logFormatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s', '%H:%M:%S')
streamHandler = logging.StreamHandler(stream=sys.stdout)
streamHandler.setFormatter(logFormatter)
rootLogger.addHandler(streamHandler)
config = Config(cmdArgs.config)
try:
config.loadConfig()
except ConfigError:
rootLogger.exception("Failed to load configuration file {}".format(cmdArgs.config))
else:
factory = DesertBotFactory(config)
reactor.run()
| Make the logging level configurable | ## Code Before:
import argparse
import logging
import os
import sys
from twisted.internet import reactor
from desertbot.config import Config, ConfigError
from desertbot.factory import DesertBotFactory
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='An IRC bot written in Python.')
parser.add_argument('-c', '--config',
help='the config file to read from',
type=str, required=True)
cmdArgs = parser.parse_args()
os.chdir(os.path.dirname(os.path.abspath(__file__)))
# Set up logging for stdout on the root 'desertbot' logger
# Modules can then just add more handlers to the root logger to capture all logs to files in various ways
rootLogger = logging.getLogger('desertbot')
rootLogger.setLevel(logging.INFO) # TODO change this from config value once it's loaded
logFormatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s', '%H:%M:%S')
streamHandler = logging.StreamHandler(stream=sys.stdout)
streamHandler.setFormatter(logFormatter)
rootLogger.addHandler(streamHandler)
config = Config(cmdArgs.config)
try:
config.loadConfig()
except ConfigError:
rootLogger.exception("Failed to load configuration file {}".format(cmdArgs.config))
else:
factory = DesertBotFactory(config)
reactor.run()
## Instruction:
Make the logging level configurable
## Code After:
import argparse
import logging
import os
import sys
from twisted.internet import reactor
from desertbot.config import Config, ConfigError
from desertbot.factory import DesertBotFactory
if __name__ == '__main__':
parser = argparse.ArgumentParser(description='An IRC bot written in Python.')
parser.add_argument('-c', '--config',
help='the config file to read from',
type=str, required=True)
parser.add_argument('-l', '--loglevel',
help='the logging level (default INFO)',
type=str, default='INFO')
cmdArgs = parser.parse_args()
os.chdir(os.path.dirname(os.path.abspath(__file__)))
# Set up logging for stdout on the root 'desertbot' logger
# Modules can then just add more handlers to the root logger to capture all logs to files in various ways
rootLogger = logging.getLogger('desertbot')
numericLevel = getattr(logging, cmdArgs.loglevel.upper(), None)
if isinstance(numericLevel, int):
rootLogger.setLevel(numericLevel)
else:
raise ValueError('Invalid log level {}'.format(cmdArgs.loglevel))
logFormatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s', '%H:%M:%S')
streamHandler = logging.StreamHandler(stream=sys.stdout)
streamHandler.setFormatter(logFormatter)
rootLogger.addHandler(streamHandler)
config = Config(cmdArgs.config)
try:
config.loadConfig()
except ConfigError:
rootLogger.exception("Failed to load configuration file {}".format(cmdArgs.config))
else:
factory = DesertBotFactory(config)
reactor.run()
|
f13fc280f25996ec7f4924647fdc879779f51737 | project/tools/normalize.py | project/tools/normalize.py |
import os
import IPython.nbformat.current as nbf
from glob import glob
from lib import get_project_dir
import sys
def normalize(in_file, out_file):
worksheet = in_file.worksheets[0]
cell_list = []
# add graphic here & append to cell_list
for cell in worksheet.cells:
if cell.cell_type == ("code"):
cell.outputs = []
cell.prompt_number = ""
cell_list.append(cell)
output_nb = nbf.new_notebook() # XXX should set name ...
output_nb.worksheets.append(nbf.new_worksheet(cells=cell_list))
nbf.write(output_nb, out_file, "ipynb")
if __name__ == "__main__":
if len(sys.argv) == 3:
infile = open(sys.argv[1])
outfile = open(sys.argv[2],"w")
else:
infile = sys.stdin
outfile = sys.stdout
normalize(nbf.read(infile, "ipynb"), sys.stdout) |
import os
import IPython.nbformat.current as nbf
from glob import glob
from lib import get_project_dir
import sys
def normalize(in_file, out_file):
worksheet = in_file.worksheets[0]
cell_list = []
# add graphic here & append to cell_list
for cell in worksheet.cells:
if cell.cell_type == ("code"):
cell.outputs = []
cell.prompt_number = ""
cell_list.append(cell)
output_nb = nbf.new_notebook() # XXX should set name ...
output_nb.worksheets.append(nbf.new_worksheet(cells=cell_list))
nbf.write(output_nb, out_file, "ipynb")
if __name__ == "__main__":
if len(sys.argv) == 3:
infile = open(sys.argv[1])
outfile = open(sys.argv[2],"w")
elif len(sys.argv) != 1:
sys.exit("normalize: two arguments or none, please")
else:
infile = sys.stdin
outfile = sys.stdout
try:
normalize(nbf.read(infile, "ipynb"), outfile)
except Exception as e:
sys.exit("Normalization error: '{}'".format(str(e))) | Allow two command arguments for in and out files, or none for standard filter operations | Allow two command arguments for in and out files, or none for standard filter operations
| Python | mit | holdenweb/nbtools,holdenweb/nbtools |
import os
import IPython.nbformat.current as nbf
from glob import glob
from lib import get_project_dir
import sys
def normalize(in_file, out_file):
worksheet = in_file.worksheets[0]
cell_list = []
# add graphic here & append to cell_list
for cell in worksheet.cells:
if cell.cell_type == ("code"):
cell.outputs = []
cell.prompt_number = ""
cell_list.append(cell)
output_nb = nbf.new_notebook() # XXX should set name ...
output_nb.worksheets.append(nbf.new_worksheet(cells=cell_list))
nbf.write(output_nb, out_file, "ipynb")
if __name__ == "__main__":
if len(sys.argv) == 3:
infile = open(sys.argv[1])
outfile = open(sys.argv[2],"w")
+ elif len(sys.argv) != 1:
+ sys.exit("normalize: two arguments or none, please")
else:
infile = sys.stdin
outfile = sys.stdout
-
+ try:
- normalize(nbf.read(infile, "ipynb"), sys.stdout)
+ normalize(nbf.read(infile, "ipynb"), outfile)
+ except Exception as e:
+ sys.exit("Normalization error: '{}'".format(str(e))) | Allow two command arguments for in and out files, or none for standard filter operations | ## Code Before:
import os
import IPython.nbformat.current as nbf
from glob import glob
from lib import get_project_dir
import sys
def normalize(in_file, out_file):
worksheet = in_file.worksheets[0]
cell_list = []
# add graphic here & append to cell_list
for cell in worksheet.cells:
if cell.cell_type == ("code"):
cell.outputs = []
cell.prompt_number = ""
cell_list.append(cell)
output_nb = nbf.new_notebook() # XXX should set name ...
output_nb.worksheets.append(nbf.new_worksheet(cells=cell_list))
nbf.write(output_nb, out_file, "ipynb")
if __name__ == "__main__":
if len(sys.argv) == 3:
infile = open(sys.argv[1])
outfile = open(sys.argv[2],"w")
else:
infile = sys.stdin
outfile = sys.stdout
normalize(nbf.read(infile, "ipynb"), sys.stdout)
## Instruction:
Allow two command arguments for in and out files, or none for standard filter operations
## Code After:
import os
import IPython.nbformat.current as nbf
from glob import glob
from lib import get_project_dir
import sys
def normalize(in_file, out_file):
worksheet = in_file.worksheets[0]
cell_list = []
# add graphic here & append to cell_list
for cell in worksheet.cells:
if cell.cell_type == ("code"):
cell.outputs = []
cell.prompt_number = ""
cell_list.append(cell)
output_nb = nbf.new_notebook() # XXX should set name ...
output_nb.worksheets.append(nbf.new_worksheet(cells=cell_list))
nbf.write(output_nb, out_file, "ipynb")
if __name__ == "__main__":
if len(sys.argv) == 3:
infile = open(sys.argv[1])
outfile = open(sys.argv[2],"w")
elif len(sys.argv) != 1:
sys.exit("normalize: two arguments or none, please")
else:
infile = sys.stdin
outfile = sys.stdout
try:
normalize(nbf.read(infile, "ipynb"), outfile)
except Exception as e:
sys.exit("Normalization error: '{}'".format(str(e))) |
7119930b662a20d9e9bbca230f8a6485efcb7c44 | flask_appconfig/middleware.py | flask_appconfig/middleware.py |
class ReverseProxied(object):
'''Wrap the application in this middleware and configure the
front-end server to add these headers, to let you quietly bind
this to a URL other than / and to an HTTP scheme that is
different than what is used locally.
In nginx:
location /myprefix {
proxy_pass http://192.168.0.1:5001;
proxy_set_header Host $host;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
proxy_set_header X-Scheme $scheme;
proxy_set_header X-Script-Name /myprefix;
}
:param app: the WSGI application
'''
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
script_name = environ.get('HTTP_X_SCRIPT_NAME', '')
if script_name:
environ['SCRIPT_NAME'] = script_name
path_info = environ['PATH_INFO']
if path_info.startswith(script_name):
environ['PATH_INFO'] = path_info[len(script_name):]
scheme = environ.get('HTTP_X_SCHEME', '')
if scheme:
environ['wsgi.url_scheme'] = scheme
return self.app(environ, start_response)
|
class ReverseProxied(object):
'''Wrap the application in this middleware and configure the
front-end server to add these headers, to let you quietly bind
this to a URL other than / and to an HTTP scheme that is
different than what is used locally.
In nginx:
location /myprefix {
proxy_pass http://192.168.0.1:5001;
proxy_set_header Host $host;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
proxy_set_header X-Scheme $scheme;
proxy_set_header X-Script-Name /myprefix;
}
:param app: the WSGI application
'''
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
script_name = environ.get('HTTP_X_SCRIPT_NAME', '')
if script_name:
environ['SCRIPT_NAME'] = script_name
path_info = environ['PATH_INFO']
if path_info.startswith(script_name):
environ['PATH_INFO'] = path_info[len(script_name):]
scheme = environ.get('HTTP_X_SCHEME', '')
if scheme:
environ['wsgi.url_scheme'] = scheme
return self.app(environ, start_response)
# pass through other attributes, like .run() when using werkzeug
def __getattr__(self, key):
return getattr(self.app, key)
| Add __getattr__ passthrough on ReverseProxied. | Add __getattr__ passthrough on ReverseProxied.
| Python | mit | mbr/flask-appconfig |
class ReverseProxied(object):
'''Wrap the application in this middleware and configure the
front-end server to add these headers, to let you quietly bind
this to a URL other than / and to an HTTP scheme that is
different than what is used locally.
In nginx:
location /myprefix {
proxy_pass http://192.168.0.1:5001;
proxy_set_header Host $host;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
proxy_set_header X-Scheme $scheme;
proxy_set_header X-Script-Name /myprefix;
}
:param app: the WSGI application
'''
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
script_name = environ.get('HTTP_X_SCRIPT_NAME', '')
if script_name:
environ['SCRIPT_NAME'] = script_name
path_info = environ['PATH_INFO']
if path_info.startswith(script_name):
environ['PATH_INFO'] = path_info[len(script_name):]
scheme = environ.get('HTTP_X_SCHEME', '')
if scheme:
environ['wsgi.url_scheme'] = scheme
return self.app(environ, start_response)
+ # pass through other attributes, like .run() when using werkzeug
+ def __getattr__(self, key):
+ return getattr(self.app, key)
+ | Add __getattr__ passthrough on ReverseProxied. | ## Code Before:
class ReverseProxied(object):
'''Wrap the application in this middleware and configure the
front-end server to add these headers, to let you quietly bind
this to a URL other than / and to an HTTP scheme that is
different than what is used locally.
In nginx:
location /myprefix {
proxy_pass http://192.168.0.1:5001;
proxy_set_header Host $host;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
proxy_set_header X-Scheme $scheme;
proxy_set_header X-Script-Name /myprefix;
}
:param app: the WSGI application
'''
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
script_name = environ.get('HTTP_X_SCRIPT_NAME', '')
if script_name:
environ['SCRIPT_NAME'] = script_name
path_info = environ['PATH_INFO']
if path_info.startswith(script_name):
environ['PATH_INFO'] = path_info[len(script_name):]
scheme = environ.get('HTTP_X_SCHEME', '')
if scheme:
environ['wsgi.url_scheme'] = scheme
return self.app(environ, start_response)
## Instruction:
Add __getattr__ passthrough on ReverseProxied.
## Code After:
class ReverseProxied(object):
'''Wrap the application in this middleware and configure the
front-end server to add these headers, to let you quietly bind
this to a URL other than / and to an HTTP scheme that is
different than what is used locally.
In nginx:
location /myprefix {
proxy_pass http://192.168.0.1:5001;
proxy_set_header Host $host;
proxy_set_header X-Forwarded-For $proxy_add_x_forwarded_for;
proxy_set_header X-Scheme $scheme;
proxy_set_header X-Script-Name /myprefix;
}
:param app: the WSGI application
'''
def __init__(self, app):
self.app = app
def __call__(self, environ, start_response):
script_name = environ.get('HTTP_X_SCRIPT_NAME', '')
if script_name:
environ['SCRIPT_NAME'] = script_name
path_info = environ['PATH_INFO']
if path_info.startswith(script_name):
environ['PATH_INFO'] = path_info[len(script_name):]
scheme = environ.get('HTTP_X_SCHEME', '')
if scheme:
environ['wsgi.url_scheme'] = scheme
return self.app(environ, start_response)
# pass through other attributes, like .run() when using werkzeug
def __getattr__(self, key):
return getattr(self.app, key)
|
ba32a22cc0cb41c4548c658a7195fab56dab6dbf | atlas/prodtask/tasks.py | atlas/prodtask/tasks.py | from __future__ import absolute_import, unicode_literals
from atlas.celerybackend.celery import app
from atlas.prestage.views import find_action_to_execute, submit_all_tapes_processed
from atlas.prodtask.hashtag import hashtag_request_to_tasks
from atlas.prodtask.mcevgen import sync_cvmfs_db
from atlas.prodtask.open_ended import check_open_ended
from atlas.prodtask.task_views import sync_old_tasks
import logging
_logger = logging.getLogger('prodtaskwebui')
@app.task
def test_celery():
_logger.info('test celery')
return 2
@app.task(ignore_result=True)
def sync_tasks():
sync_old_tasks(-1)
return None
@app.task(ignore_result=True)
def step_actions():
find_action_to_execute()
return None
@app.task(ignore_result=True)
def data_carousel():
submit_all_tapes_processed()
return None
@app.task(ignore_result=True)
def open_ended():
check_open_ended()
return None
@app.task(ignore_result=True)
def request_hashtags():
hashtag_request_to_tasks()
return None
@app.task(ignore_result=True)
def sync_evgen_jo():
sync_cvmfs_db()
return None
| from __future__ import absolute_import, unicode_literals
from atlas.celerybackend.celery import app
from atlas.prestage.views import find_action_to_execute, submit_all_tapes_processed, delete_done_staging_rules
from atlas.prodtask.hashtag import hashtag_request_to_tasks
from atlas.prodtask.mcevgen import sync_cvmfs_db
from atlas.prodtask.open_ended import check_open_ended
from atlas.prodtask.task_views import sync_old_tasks
import logging
_logger = logging.getLogger('prodtaskwebui')
@app.task
def test_celery():
_logger.info('test celery')
return 2
@app.task(ignore_result=True)
def sync_tasks():
sync_old_tasks(-1)
return None
@app.task(ignore_result=True)
def step_actions():
find_action_to_execute()
return None
@app.task(ignore_result=True)
def data_carousel():
submit_all_tapes_processed()
return None
@app.task(ignore_result=True)
def open_ended():
check_open_ended()
return None
@app.task(ignore_result=True)
def request_hashtags():
hashtag_request_to_tasks()
return None
@app.task(ignore_result=True)
def sync_evgen_jo():
sync_cvmfs_db()
return None
@app.task(ignore_result=True)
def remove_done_staging(production_requests):
delete_done_staging_rules(production_requests)
return None | Add remove done staged rules | Add remove done staged rules
| Python | apache-2.0 | PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas | from __future__ import absolute_import, unicode_literals
from atlas.celerybackend.celery import app
- from atlas.prestage.views import find_action_to_execute, submit_all_tapes_processed
+ from atlas.prestage.views import find_action_to_execute, submit_all_tapes_processed, delete_done_staging_rules
from atlas.prodtask.hashtag import hashtag_request_to_tasks
from atlas.prodtask.mcevgen import sync_cvmfs_db
from atlas.prodtask.open_ended import check_open_ended
from atlas.prodtask.task_views import sync_old_tasks
import logging
_logger = logging.getLogger('prodtaskwebui')
@app.task
def test_celery():
_logger.info('test celery')
return 2
@app.task(ignore_result=True)
def sync_tasks():
sync_old_tasks(-1)
return None
@app.task(ignore_result=True)
def step_actions():
find_action_to_execute()
return None
@app.task(ignore_result=True)
def data_carousel():
submit_all_tapes_processed()
return None
@app.task(ignore_result=True)
def open_ended():
check_open_ended()
return None
@app.task(ignore_result=True)
def request_hashtags():
hashtag_request_to_tasks()
return None
@app.task(ignore_result=True)
def sync_evgen_jo():
sync_cvmfs_db()
return None
+ @app.task(ignore_result=True)
+ def remove_done_staging(production_requests):
+ delete_done_staging_rules(production_requests)
+ return None | Add remove done staged rules | ## Code Before:
from __future__ import absolute_import, unicode_literals
from atlas.celerybackend.celery import app
from atlas.prestage.views import find_action_to_execute, submit_all_tapes_processed
from atlas.prodtask.hashtag import hashtag_request_to_tasks
from atlas.prodtask.mcevgen import sync_cvmfs_db
from atlas.prodtask.open_ended import check_open_ended
from atlas.prodtask.task_views import sync_old_tasks
import logging
_logger = logging.getLogger('prodtaskwebui')
@app.task
def test_celery():
_logger.info('test celery')
return 2
@app.task(ignore_result=True)
def sync_tasks():
sync_old_tasks(-1)
return None
@app.task(ignore_result=True)
def step_actions():
find_action_to_execute()
return None
@app.task(ignore_result=True)
def data_carousel():
submit_all_tapes_processed()
return None
@app.task(ignore_result=True)
def open_ended():
check_open_ended()
return None
@app.task(ignore_result=True)
def request_hashtags():
hashtag_request_to_tasks()
return None
@app.task(ignore_result=True)
def sync_evgen_jo():
sync_cvmfs_db()
return None
## Instruction:
Add remove done staged rules
## Code After:
from __future__ import absolute_import, unicode_literals
from atlas.celerybackend.celery import app
from atlas.prestage.views import find_action_to_execute, submit_all_tapes_processed, delete_done_staging_rules
from atlas.prodtask.hashtag import hashtag_request_to_tasks
from atlas.prodtask.mcevgen import sync_cvmfs_db
from atlas.prodtask.open_ended import check_open_ended
from atlas.prodtask.task_views import sync_old_tasks
import logging
_logger = logging.getLogger('prodtaskwebui')
@app.task
def test_celery():
_logger.info('test celery')
return 2
@app.task(ignore_result=True)
def sync_tasks():
sync_old_tasks(-1)
return None
@app.task(ignore_result=True)
def step_actions():
find_action_to_execute()
return None
@app.task(ignore_result=True)
def data_carousel():
submit_all_tapes_processed()
return None
@app.task(ignore_result=True)
def open_ended():
check_open_ended()
return None
@app.task(ignore_result=True)
def request_hashtags():
hashtag_request_to_tasks()
return None
@app.task(ignore_result=True)
def sync_evgen_jo():
sync_cvmfs_db()
return None
@app.task(ignore_result=True)
def remove_done_staging(production_requests):
delete_done_staging_rules(production_requests)
return None |
eede55d9cd39c68ef03091614096e51d7df01336 | test_scraper.py | test_scraper.py | from scraper import search_CL
from scraper import read_search_results
from scraper import parse_source
from scraper import extract_listings
import bs4
def test_search_CL():
test_body, test_encoding = search_CL(minAsk=100, maxAsk=100)
assert "<span class=\"desktop\">craigslist</span>" in test_body
assert test_encoding == 'utf-8'
def test_read_search_result():
test_body, test_encoding = read_search_results()
assert "<span class=\"desktop\">craigslist</span>" in test_body
assert test_encoding == 'utf-8'
def test_parse_source():
test_body, test_encoding = read_search_results()
test_parse = parse_source(test_body, test_encoding)
assert isinstance(test_parse, bs4.BeautifulSoup)
def test_extract_listings():
test_body, test_encoding = read_search_results()
test_parse = parse_source(test_body, test_encoding)
for row in extract_listings(test_parse):
print type(row)
assert isinstance(row, bs4.element.Tag)
| from scraper import search_CL
from scraper import read_search_results
from scraper import parse_source
from scraper import extract_listings
import bs4
def test_search_CL():
test_body, test_encoding = search_CL(minAsk=100, maxAsk=100)
assert "<span class=\"desktop\">craigslist</span>" in test_body
assert test_encoding == 'utf-8'
def test_read_search_result():
test_body, test_encoding = read_search_results()
assert "<span class=\"desktop\">craigslist</span>" in test_body
assert test_encoding == 'utf-8'
def test_parse_source():
test_body, test_encoding = read_search_results()
test_parse = parse_source(test_body, test_encoding)
assert isinstance(test_parse, bs4.BeautifulSoup)
def test_extract_listings():
test_body, test_encoding = read_search_results()
test_parse = parse_source(test_body, test_encoding)
test_data = extract_listings(test_parse)
assert isinstance(test_data, list)
for dict_ in test_data:
assert isinstance(dict_, dict) | Modify test_extract_listings() to account for the change in output from extract_listings() | Modify test_extract_listings() to account for the change in output from extract_listings()
| Python | mit | jefrailey/basic-scraper | from scraper import search_CL
from scraper import read_search_results
from scraper import parse_source
from scraper import extract_listings
import bs4
def test_search_CL():
test_body, test_encoding = search_CL(minAsk=100, maxAsk=100)
assert "<span class=\"desktop\">craigslist</span>" in test_body
assert test_encoding == 'utf-8'
def test_read_search_result():
test_body, test_encoding = read_search_results()
assert "<span class=\"desktop\">craigslist</span>" in test_body
assert test_encoding == 'utf-8'
def test_parse_source():
test_body, test_encoding = read_search_results()
test_parse = parse_source(test_body, test_encoding)
assert isinstance(test_parse, bs4.BeautifulSoup)
def test_extract_listings():
test_body, test_encoding = read_search_results()
test_parse = parse_source(test_body, test_encoding)
- for row in extract_listings(test_parse):
+ test_data = extract_listings(test_parse)
- print type(row)
- assert isinstance(row, bs4.element.Tag)
-
+ assert isinstance(test_data, list)
+ for dict_ in test_data:
+ assert isinstance(dict_, dict) | Modify test_extract_listings() to account for the change in output from extract_listings() | ## Code Before:
from scraper import search_CL
from scraper import read_search_results
from scraper import parse_source
from scraper import extract_listings
import bs4
def test_search_CL():
test_body, test_encoding = search_CL(minAsk=100, maxAsk=100)
assert "<span class=\"desktop\">craigslist</span>" in test_body
assert test_encoding == 'utf-8'
def test_read_search_result():
test_body, test_encoding = read_search_results()
assert "<span class=\"desktop\">craigslist</span>" in test_body
assert test_encoding == 'utf-8'
def test_parse_source():
test_body, test_encoding = read_search_results()
test_parse = parse_source(test_body, test_encoding)
assert isinstance(test_parse, bs4.BeautifulSoup)
def test_extract_listings():
test_body, test_encoding = read_search_results()
test_parse = parse_source(test_body, test_encoding)
for row in extract_listings(test_parse):
print type(row)
assert isinstance(row, bs4.element.Tag)
## Instruction:
Modify test_extract_listings() to account for the change in output from extract_listings()
## Code After:
from scraper import search_CL
from scraper import read_search_results
from scraper import parse_source
from scraper import extract_listings
import bs4
def test_search_CL():
test_body, test_encoding = search_CL(minAsk=100, maxAsk=100)
assert "<span class=\"desktop\">craigslist</span>" in test_body
assert test_encoding == 'utf-8'
def test_read_search_result():
test_body, test_encoding = read_search_results()
assert "<span class=\"desktop\">craigslist</span>" in test_body
assert test_encoding == 'utf-8'
def test_parse_source():
test_body, test_encoding = read_search_results()
test_parse = parse_source(test_body, test_encoding)
assert isinstance(test_parse, bs4.BeautifulSoup)
def test_extract_listings():
test_body, test_encoding = read_search_results()
test_parse = parse_source(test_body, test_encoding)
test_data = extract_listings(test_parse)
assert isinstance(test_data, list)
for dict_ in test_data:
assert isinstance(dict_, dict) |
b823233978f70d8e34a3653b309ee43b4b1e0c0d | fuel/transformers/defaults.py | fuel/transformers/defaults.py | """Commonly-used default transformers."""
from fuel.transformers import ScaleAndShift, Cast, SourcewiseTransformer
from fuel.transformers.image import ImagesFromBytes
def uint8_pixels_to_floatX(which_sources):
return (
(ScaleAndShift, [1 / 255.0, 0], {'which_sources': which_sources}),
(Cast, ['floatX'], {'which_sources': which_sources}))
class ToBytes(SourcewiseTransformer):
"""Transform a stream of ndarray examples to bytes.
Notes
-----
Used for retrieving variable-length byte data stored as, e.g. a uint8
ragged array.
"""
def __init__(self, stream, **kwargs):
kwargs.setdefault('produces_examples', stream.produces_examples)
axis_labels = stream.axis_labels
for source in kwargs.get('which_sources', stream.sources):
axis_labels[source] = (('batch', 'bytes')
if 'batch' in axis_labels.get(source, ())
else ('bytes',))
kwargs.setdefault('axis_labels', axis_labels)
super(ToBytes, self).__init__(stream, **kwargs)
def transform_source_example(self, example, _):
return example.tostring()
def transform_source_batch(self, batch, _):
return [example.tostring() for example in batch]
def rgb_images_from_encoded_bytes(which_sources):
return ((ToBytes, [], {'which_sources': ('encoded_images',)}),
(ImagesFromBytes, [], {'which_sources': ('encoded_images',)}))
| """Commonly-used default transformers."""
from fuel.transformers import ScaleAndShift, Cast, SourcewiseTransformer
from fuel.transformers.image import ImagesFromBytes
def uint8_pixels_to_floatX(which_sources):
return (
(ScaleAndShift, [1 / 255.0, 0], {'which_sources': which_sources}),
(Cast, ['floatX'], {'which_sources': which_sources}))
class ToBytes(SourcewiseTransformer):
"""Transform a stream of ndarray examples to bytes.
Notes
-----
Used for retrieving variable-length byte data stored as, e.g. a uint8
ragged array.
"""
def __init__(self, stream, **kwargs):
kwargs.setdefault('produces_examples', stream.produces_examples)
axis_labels = (stream.axis_labels
if stream.axis_labels is not None
else {})
for source in kwargs.get('which_sources', stream.sources):
axis_labels[source] = (('batch', 'bytes')
if 'batch' in axis_labels.get(source, ())
else ('bytes',))
kwargs.setdefault('axis_labels', axis_labels)
super(ToBytes, self).__init__(stream, **kwargs)
def transform_source_example(self, example, _):
return example.tostring()
def transform_source_batch(self, batch, _):
return [example.tostring() for example in batch]
def rgb_images_from_encoded_bytes(which_sources):
return ((ToBytes, [], {'which_sources': ('encoded_images',)}),
(ImagesFromBytes, [], {'which_sources': ('encoded_images',)}))
| Handle None axis_labels in ToBytes. | Handle None axis_labels in ToBytes.
| Python | mit | udibr/fuel,markusnagel/fuel,vdumoulin/fuel,mila-udem/fuel,dmitriy-serdyuk/fuel,udibr/fuel,markusnagel/fuel,aalmah/fuel,vdumoulin/fuel,aalmah/fuel,capybaralet/fuel,janchorowski/fuel,mila-udem/fuel,dribnet/fuel,capybaralet/fuel,dmitriy-serdyuk/fuel,janchorowski/fuel,dribnet/fuel | """Commonly-used default transformers."""
from fuel.transformers import ScaleAndShift, Cast, SourcewiseTransformer
from fuel.transformers.image import ImagesFromBytes
def uint8_pixels_to_floatX(which_sources):
return (
(ScaleAndShift, [1 / 255.0, 0], {'which_sources': which_sources}),
(Cast, ['floatX'], {'which_sources': which_sources}))
class ToBytes(SourcewiseTransformer):
"""Transform a stream of ndarray examples to bytes.
Notes
-----
Used for retrieving variable-length byte data stored as, e.g. a uint8
ragged array.
"""
def __init__(self, stream, **kwargs):
kwargs.setdefault('produces_examples', stream.produces_examples)
- axis_labels = stream.axis_labels
+ axis_labels = (stream.axis_labels
+ if stream.axis_labels is not None
+ else {})
for source in kwargs.get('which_sources', stream.sources):
axis_labels[source] = (('batch', 'bytes')
if 'batch' in axis_labels.get(source, ())
else ('bytes',))
kwargs.setdefault('axis_labels', axis_labels)
super(ToBytes, self).__init__(stream, **kwargs)
def transform_source_example(self, example, _):
return example.tostring()
def transform_source_batch(self, batch, _):
return [example.tostring() for example in batch]
def rgb_images_from_encoded_bytes(which_sources):
return ((ToBytes, [], {'which_sources': ('encoded_images',)}),
(ImagesFromBytes, [], {'which_sources': ('encoded_images',)}))
| Handle None axis_labels in ToBytes. | ## Code Before:
"""Commonly-used default transformers."""
from fuel.transformers import ScaleAndShift, Cast, SourcewiseTransformer
from fuel.transformers.image import ImagesFromBytes
def uint8_pixels_to_floatX(which_sources):
return (
(ScaleAndShift, [1 / 255.0, 0], {'which_sources': which_sources}),
(Cast, ['floatX'], {'which_sources': which_sources}))
class ToBytes(SourcewiseTransformer):
"""Transform a stream of ndarray examples to bytes.
Notes
-----
Used for retrieving variable-length byte data stored as, e.g. a uint8
ragged array.
"""
def __init__(self, stream, **kwargs):
kwargs.setdefault('produces_examples', stream.produces_examples)
axis_labels = stream.axis_labels
for source in kwargs.get('which_sources', stream.sources):
axis_labels[source] = (('batch', 'bytes')
if 'batch' in axis_labels.get(source, ())
else ('bytes',))
kwargs.setdefault('axis_labels', axis_labels)
super(ToBytes, self).__init__(stream, **kwargs)
def transform_source_example(self, example, _):
return example.tostring()
def transform_source_batch(self, batch, _):
return [example.tostring() for example in batch]
def rgb_images_from_encoded_bytes(which_sources):
return ((ToBytes, [], {'which_sources': ('encoded_images',)}),
(ImagesFromBytes, [], {'which_sources': ('encoded_images',)}))
## Instruction:
Handle None axis_labels in ToBytes.
## Code After:
"""Commonly-used default transformers."""
from fuel.transformers import ScaleAndShift, Cast, SourcewiseTransformer
from fuel.transformers.image import ImagesFromBytes
def uint8_pixels_to_floatX(which_sources):
return (
(ScaleAndShift, [1 / 255.0, 0], {'which_sources': which_sources}),
(Cast, ['floatX'], {'which_sources': which_sources}))
class ToBytes(SourcewiseTransformer):
"""Transform a stream of ndarray examples to bytes.
Notes
-----
Used for retrieving variable-length byte data stored as, e.g. a uint8
ragged array.
"""
def __init__(self, stream, **kwargs):
kwargs.setdefault('produces_examples', stream.produces_examples)
axis_labels = (stream.axis_labels
if stream.axis_labels is not None
else {})
for source in kwargs.get('which_sources', stream.sources):
axis_labels[source] = (('batch', 'bytes')
if 'batch' in axis_labels.get(source, ())
else ('bytes',))
kwargs.setdefault('axis_labels', axis_labels)
super(ToBytes, self).__init__(stream, **kwargs)
def transform_source_example(self, example, _):
return example.tostring()
def transform_source_batch(self, batch, _):
return [example.tostring() for example in batch]
def rgb_images_from_encoded_bytes(which_sources):
return ((ToBytes, [], {'which_sources': ('encoded_images',)}),
(ImagesFromBytes, [], {'which_sources': ('encoded_images',)}))
|
beeae2daf35da275d5f9e1ad01516c917319bf00 | gapipy/resources/geo/state.py | gapipy/resources/geo/state.py | from __future__ import unicode_literals
from ..base import Resource
from ...utils import enforce_string_type
class State(Resource):
_resource_name = 'states'
_as_is_fields = ['id', 'href', 'name']
_resource_fields = [('country', 'Country')]
@enforce_string_type
def __repr__(self):
return '<{}: {}>'.format(self.__class__.__name__, self.name)
| from __future__ import unicode_literals
from ..base import Resource
from ...utils import enforce_string_type
class State(Resource):
_resource_name = 'states'
_as_is_fields = ['id', 'href', 'name']
_resource_fields = [
('country', 'Country'),
('place', 'Place'),
]
@enforce_string_type
def __repr__(self):
return '<{}: {}>'.format(self.__class__.__name__, self.name)
| Add Place reference to State model | Add Place reference to State model
| Python | mit | gadventures/gapipy | from __future__ import unicode_literals
from ..base import Resource
from ...utils import enforce_string_type
class State(Resource):
_resource_name = 'states'
_as_is_fields = ['id', 'href', 'name']
- _resource_fields = [('country', 'Country')]
+ _resource_fields = [
+ ('country', 'Country'),
+ ('place', 'Place'),
+ ]
@enforce_string_type
def __repr__(self):
return '<{}: {}>'.format(self.__class__.__name__, self.name)
| Add Place reference to State model | ## Code Before:
from __future__ import unicode_literals
from ..base import Resource
from ...utils import enforce_string_type
class State(Resource):
_resource_name = 'states'
_as_is_fields = ['id', 'href', 'name']
_resource_fields = [('country', 'Country')]
@enforce_string_type
def __repr__(self):
return '<{}: {}>'.format(self.__class__.__name__, self.name)
## Instruction:
Add Place reference to State model
## Code After:
from __future__ import unicode_literals
from ..base import Resource
from ...utils import enforce_string_type
class State(Resource):
_resource_name = 'states'
_as_is_fields = ['id', 'href', 'name']
_resource_fields = [
('country', 'Country'),
('place', 'Place'),
]
@enforce_string_type
def __repr__(self):
return '<{}: {}>'.format(self.__class__.__name__, self.name)
|
cedac36d38ff0bf70abc1c9193948a288e858a01 | kitsune/lib/pipeline_compilers.py | kitsune/lib/pipeline_compilers.py | import re
from django.conf import settings
from django.utils.encoding import smart_bytes
from pipeline.compilers import CompilerBase
from pipeline.exceptions import CompilerError
class BrowserifyCompiler(CompilerBase):
output_extension = 'browserified.js'
def match_file(self, path):
# Allow for cache busting hashes between ".browserify" and ".js"
return re.search(r'\.browserify(\.[a-fA-F0-9]+)?\.js$', path) is not None
def compile_file(self, infile, outfile, outdated=False, force=False):
command = "%s %s %s > %s" % (
getattr(settings, 'PIPELINE_BROWSERIFY_BINARY', '/usr/bin/env browserify'),
getattr(settings, 'PIPELINE_BROWSERIFY_ARGUMENTS', ''),
infile,
outfile
)
return self.execute_command(command)
def execute_command(self, command, content=None, cwd=None):
"""This is like the one in SubProcessCompiler, except it checks the exit code."""
import subprocess
pipe = subprocess.Popen(command, shell=True, cwd=cwd,
stdout=subprocess.PIPE, stdin=subprocess.PIPE,
stderr=subprocess.PIPE)
if content:
content = smart_bytes(content)
stdout, stderr = pipe.communicate(content)
if self.verbose:
print(stderr)
if pipe.returncode != 0:
raise CompilerError(stderr)
return stdout
| import re
from django.conf import settings
from django.utils.encoding import smart_bytes
from pipeline.compilers import CompilerBase
from pipeline.exceptions import CompilerError
class BrowserifyCompiler(CompilerBase):
output_extension = 'browserified.js'
def match_file(self, path):
# Allow for cache busting hashes between ".browserify" and ".js"
return re.search(r'\.browserify(\.[a-fA-F0-9]+)?\.js$', path) is not None
def compile_file(self, infile, outfile, outdated=False, force=False):
pipeline_settings = getattr(settings, 'PIPELINE', {})
command = "%s %s %s > %s" % (
pipeline_settings.get('BROWSERIFY_BINARY', '/usr/bin/env browserify'),
pipeline_settings.get('BROWSERIFY_ARGUMENTS', ''),
infile,
outfile
)
return self.execute_command(command)
def execute_command(self, command, content=None, cwd=None):
"""This is like the one in SubProcessCompiler, except it checks the exit code."""
import subprocess
pipe = subprocess.Popen(command, shell=True, cwd=cwd,
stdout=subprocess.PIPE, stdin=subprocess.PIPE,
stderr=subprocess.PIPE)
if content:
content = smart_bytes(content)
stdout, stderr = pipe.communicate(content)
if self.verbose:
print(stderr)
if pipe.returncode != 0:
raise CompilerError(stderr)
return stdout
| Update BrowserifyCompiler for n Pipeline settings. | Update BrowserifyCompiler for n Pipeline settings.
| Python | bsd-3-clause | mythmon/kitsune,MikkCZ/kitsune,brittanystoroz/kitsune,anushbmx/kitsune,MikkCZ/kitsune,anushbmx/kitsune,safwanrahman/kitsune,brittanystoroz/kitsune,MikkCZ/kitsune,mozilla/kitsune,safwanrahman/kitsune,mythmon/kitsune,anushbmx/kitsune,mythmon/kitsune,safwanrahman/kitsune,mozilla/kitsune,brittanystoroz/kitsune,mythmon/kitsune,mozilla/kitsune,anushbmx/kitsune,MikkCZ/kitsune,mozilla/kitsune,brittanystoroz/kitsune,safwanrahman/kitsune | import re
from django.conf import settings
from django.utils.encoding import smart_bytes
from pipeline.compilers import CompilerBase
from pipeline.exceptions import CompilerError
class BrowserifyCompiler(CompilerBase):
output_extension = 'browserified.js'
def match_file(self, path):
# Allow for cache busting hashes between ".browserify" and ".js"
return re.search(r'\.browserify(\.[a-fA-F0-9]+)?\.js$', path) is not None
def compile_file(self, infile, outfile, outdated=False, force=False):
+ pipeline_settings = getattr(settings, 'PIPELINE', {})
command = "%s %s %s > %s" % (
- getattr(settings, 'PIPELINE_BROWSERIFY_BINARY', '/usr/bin/env browserify'),
+ pipeline_settings.get('BROWSERIFY_BINARY', '/usr/bin/env browserify'),
- getattr(settings, 'PIPELINE_BROWSERIFY_ARGUMENTS', ''),
+ pipeline_settings.get('BROWSERIFY_ARGUMENTS', ''),
infile,
outfile
)
return self.execute_command(command)
def execute_command(self, command, content=None, cwd=None):
"""This is like the one in SubProcessCompiler, except it checks the exit code."""
import subprocess
pipe = subprocess.Popen(command, shell=True, cwd=cwd,
stdout=subprocess.PIPE, stdin=subprocess.PIPE,
stderr=subprocess.PIPE)
if content:
content = smart_bytes(content)
stdout, stderr = pipe.communicate(content)
if self.verbose:
print(stderr)
if pipe.returncode != 0:
raise CompilerError(stderr)
return stdout
| Update BrowserifyCompiler for n Pipeline settings. | ## Code Before:
import re
from django.conf import settings
from django.utils.encoding import smart_bytes
from pipeline.compilers import CompilerBase
from pipeline.exceptions import CompilerError
class BrowserifyCompiler(CompilerBase):
output_extension = 'browserified.js'
def match_file(self, path):
# Allow for cache busting hashes between ".browserify" and ".js"
return re.search(r'\.browserify(\.[a-fA-F0-9]+)?\.js$', path) is not None
def compile_file(self, infile, outfile, outdated=False, force=False):
command = "%s %s %s > %s" % (
getattr(settings, 'PIPELINE_BROWSERIFY_BINARY', '/usr/bin/env browserify'),
getattr(settings, 'PIPELINE_BROWSERIFY_ARGUMENTS', ''),
infile,
outfile
)
return self.execute_command(command)
def execute_command(self, command, content=None, cwd=None):
"""This is like the one in SubProcessCompiler, except it checks the exit code."""
import subprocess
pipe = subprocess.Popen(command, shell=True, cwd=cwd,
stdout=subprocess.PIPE, stdin=subprocess.PIPE,
stderr=subprocess.PIPE)
if content:
content = smart_bytes(content)
stdout, stderr = pipe.communicate(content)
if self.verbose:
print(stderr)
if pipe.returncode != 0:
raise CompilerError(stderr)
return stdout
## Instruction:
Update BrowserifyCompiler for n Pipeline settings.
## Code After:
import re
from django.conf import settings
from django.utils.encoding import smart_bytes
from pipeline.compilers import CompilerBase
from pipeline.exceptions import CompilerError
class BrowserifyCompiler(CompilerBase):
output_extension = 'browserified.js'
def match_file(self, path):
# Allow for cache busting hashes between ".browserify" and ".js"
return re.search(r'\.browserify(\.[a-fA-F0-9]+)?\.js$', path) is not None
def compile_file(self, infile, outfile, outdated=False, force=False):
pipeline_settings = getattr(settings, 'PIPELINE', {})
command = "%s %s %s > %s" % (
pipeline_settings.get('BROWSERIFY_BINARY', '/usr/bin/env browserify'),
pipeline_settings.get('BROWSERIFY_ARGUMENTS', ''),
infile,
outfile
)
return self.execute_command(command)
def execute_command(self, command, content=None, cwd=None):
"""This is like the one in SubProcessCompiler, except it checks the exit code."""
import subprocess
pipe = subprocess.Popen(command, shell=True, cwd=cwd,
stdout=subprocess.PIPE, stdin=subprocess.PIPE,
stderr=subprocess.PIPE)
if content:
content = smart_bytes(content)
stdout, stderr = pipe.communicate(content)
if self.verbose:
print(stderr)
if pipe.returncode != 0:
raise CompilerError(stderr)
return stdout
|
c1f8d5817b8c94b422c0d454dcc0fa3c00e751b6 | activelink/tests/urls.py | activelink/tests/urls.py | from django import VERSION as DJANGO_VERSION
from django.http import HttpResponse
if DJANGO_VERSION >= (1, 6):
from django.conf.urls import patterns, url
else:
from django.conf.urls.defaults import patterns, url
urlpatterns = patterns('',
url(r'^test-url/$', lambda r: HttpResponse('ok'), name='test'),
url(r'^test-url-with-arg/([-\w]+)/$', lambda r, arg: HttpResponse('ok'), name='test_with_arg'),
url(r'^test-url-with-kwarg/(?P<arg>[-\w]+)/$', lambda r, arg: HttpResponse('ok'), name='test_with_kwarg'),
)
| from django import VERSION as DJANGO_VERSION
from django.http import HttpResponse
if DJANGO_VERSION >= (1, 10):
from django.conf.urls import url
elif DJANGO_VERSION >= (1, 6):
from django.conf.urls import patterns, url
else:
from django.conf.urls.defaults import patterns, url
urlpatterns = [
url(r'^test-url/$', lambda r: HttpResponse('ok'), name='test'),
url(r'^test-url-with-arg/([-\w]+)/$', lambda r, arg: HttpResponse('ok'), name='test_with_arg'),
url(r'^test-url-with-kwarg/(?P<arg>[-\w]+)/$', lambda r, arg: HttpResponse('ok'), name='test_with_kwarg'),
]
if DJANGO_VERSION < (1, 10):
urlpatterns = patterns('', *urlpatterns)
| Add support for Django 1.11 | Add support for Django 1.11
| Python | unlicense | j4mie/django-activelink | from django import VERSION as DJANGO_VERSION
from django.http import HttpResponse
+
+ if DJANGO_VERSION >= (1, 10):
+ from django.conf.urls import url
- if DJANGO_VERSION >= (1, 6):
+ elif DJANGO_VERSION >= (1, 6):
from django.conf.urls import patterns, url
else:
from django.conf.urls.defaults import patterns, url
- urlpatterns = patterns('',
+ urlpatterns = [
url(r'^test-url/$', lambda r: HttpResponse('ok'), name='test'),
url(r'^test-url-with-arg/([-\w]+)/$', lambda r, arg: HttpResponse('ok'), name='test_with_arg'),
url(r'^test-url-with-kwarg/(?P<arg>[-\w]+)/$', lambda r, arg: HttpResponse('ok'), name='test_with_kwarg'),
- )
+ ]
+ if DJANGO_VERSION < (1, 10):
+ urlpatterns = patterns('', *urlpatterns)
+ | Add support for Django 1.11 | ## Code Before:
from django import VERSION as DJANGO_VERSION
from django.http import HttpResponse
if DJANGO_VERSION >= (1, 6):
from django.conf.urls import patterns, url
else:
from django.conf.urls.defaults import patterns, url
urlpatterns = patterns('',
url(r'^test-url/$', lambda r: HttpResponse('ok'), name='test'),
url(r'^test-url-with-arg/([-\w]+)/$', lambda r, arg: HttpResponse('ok'), name='test_with_arg'),
url(r'^test-url-with-kwarg/(?P<arg>[-\w]+)/$', lambda r, arg: HttpResponse('ok'), name='test_with_kwarg'),
)
## Instruction:
Add support for Django 1.11
## Code After:
from django import VERSION as DJANGO_VERSION
from django.http import HttpResponse
if DJANGO_VERSION >= (1, 10):
from django.conf.urls import url
elif DJANGO_VERSION >= (1, 6):
from django.conf.urls import patterns, url
else:
from django.conf.urls.defaults import patterns, url
urlpatterns = [
url(r'^test-url/$', lambda r: HttpResponse('ok'), name='test'),
url(r'^test-url-with-arg/([-\w]+)/$', lambda r, arg: HttpResponse('ok'), name='test_with_arg'),
url(r'^test-url-with-kwarg/(?P<arg>[-\w]+)/$', lambda r, arg: HttpResponse('ok'), name='test_with_kwarg'),
]
if DJANGO_VERSION < (1, 10):
urlpatterns = patterns('', *urlpatterns)
|
c2ca03ba94349340447a316ff21bcb26631e308f | lms/djangoapps/discussion/settings/common.py | lms/djangoapps/discussion/settings/common.py | """Common environment variables unique to the discussion plugin."""
def plugin_settings(settings):
"""Settings for the discussions plugin. """
settings.FEATURES['ALLOW_HIDING_DISCUSSION_TAB'] = False
settings.DISCUSSION_SETTINGS = {
'MAX_COMMENT_DEPTH': 2,
'COURSE_PUBLISH_TASK_DELAY': 30,
}
| """Common environment variables unique to the discussion plugin."""
def plugin_settings(settings):
"""Settings for the discussions plugin. """
# .. toggle_name: ALLOW_HIDING_DISCUSSION_TAB
# .. toggle_implementation: DjangoSetting
# .. toggle_default: False
# .. toggle_description: If True, it adds an option to show/hide the discussions tab.
# .. toggle_use_cases: open_edx
# .. toggle_creation_date: 2015-06-15
# .. toggle_target_removal_date: None
# .. toggle_warnings: None
# .. toggle_tickets: https://github.com/edx/edx-platform/pull/8474
settings.FEATURES['ALLOW_HIDING_DISCUSSION_TAB'] = False
settings.DISCUSSION_SETTINGS = {
'MAX_COMMENT_DEPTH': 2,
'COURSE_PUBLISH_TASK_DELAY': 30,
}
| Add annotation for ALLOW_HIDING_DISCUSSION_TAB feature flag | Add annotation for ALLOW_HIDING_DISCUSSION_TAB feature flag
| Python | agpl-3.0 | EDUlib/edx-platform,eduNEXT/edx-platform,EDUlib/edx-platform,eduNEXT/edunext-platform,edx/edx-platform,eduNEXT/edx-platform,angelapper/edx-platform,eduNEXT/edx-platform,eduNEXT/edx-platform,arbrandes/edx-platform,arbrandes/edx-platform,eduNEXT/edunext-platform,EDUlib/edx-platform,edx/edx-platform,angelapper/edx-platform,edx/edx-platform,angelapper/edx-platform,eduNEXT/edunext-platform,arbrandes/edx-platform,edx/edx-platform,EDUlib/edx-platform,angelapper/edx-platform,arbrandes/edx-platform,eduNEXT/edunext-platform | """Common environment variables unique to the discussion plugin."""
def plugin_settings(settings):
"""Settings for the discussions plugin. """
+ # .. toggle_name: ALLOW_HIDING_DISCUSSION_TAB
+ # .. toggle_implementation: DjangoSetting
+ # .. toggle_default: False
+ # .. toggle_description: If True, it adds an option to show/hide the discussions tab.
+ # .. toggle_use_cases: open_edx
+ # .. toggle_creation_date: 2015-06-15
+ # .. toggle_target_removal_date: None
+ # .. toggle_warnings: None
+ # .. toggle_tickets: https://github.com/edx/edx-platform/pull/8474
settings.FEATURES['ALLOW_HIDING_DISCUSSION_TAB'] = False
settings.DISCUSSION_SETTINGS = {
'MAX_COMMENT_DEPTH': 2,
'COURSE_PUBLISH_TASK_DELAY': 30,
}
| Add annotation for ALLOW_HIDING_DISCUSSION_TAB feature flag | ## Code Before:
"""Common environment variables unique to the discussion plugin."""
def plugin_settings(settings):
"""Settings for the discussions plugin. """
settings.FEATURES['ALLOW_HIDING_DISCUSSION_TAB'] = False
settings.DISCUSSION_SETTINGS = {
'MAX_COMMENT_DEPTH': 2,
'COURSE_PUBLISH_TASK_DELAY': 30,
}
## Instruction:
Add annotation for ALLOW_HIDING_DISCUSSION_TAB feature flag
## Code After:
"""Common environment variables unique to the discussion plugin."""
def plugin_settings(settings):
"""Settings for the discussions plugin. """
# .. toggle_name: ALLOW_HIDING_DISCUSSION_TAB
# .. toggle_implementation: DjangoSetting
# .. toggle_default: False
# .. toggle_description: If True, it adds an option to show/hide the discussions tab.
# .. toggle_use_cases: open_edx
# .. toggle_creation_date: 2015-06-15
# .. toggle_target_removal_date: None
# .. toggle_warnings: None
# .. toggle_tickets: https://github.com/edx/edx-platform/pull/8474
settings.FEATURES['ALLOW_HIDING_DISCUSSION_TAB'] = False
settings.DISCUSSION_SETTINGS = {
'MAX_COMMENT_DEPTH': 2,
'COURSE_PUBLISH_TASK_DELAY': 30,
}
|
90ca340883077f57ba63127db058a8d244ec6f4c | molecule/ui/tests/conftest.py | molecule/ui/tests/conftest.py | import pytest
from selenium import webdriver
from selenium.webdriver.chrome.options import Options
from selenium.webdriver.common.keys import Keys
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.support import expected_conditions
import time
from webdriver_manager.chrome import ChromeDriverManager
from webdriver_manager.utils import ChromeType
@pytest.fixture(scope="session")
def chromedriver():
try:
options = Options()
options.headless = True
options.add_argument('--no-sandbox')
options.add_argument('--disable-dev-shm-usage')
options.add_argument("--disable-gpu")
driver = webdriver.Chrome(ChromeDriverManager(chrome_type=ChromeType.CHROMIUM).install(), options=options)
url = 'http://localhost:9000'
driver.get(url + "/gettingstarted")
WebDriverWait(driver, 30).until(expected_conditions.title_contains('Sign in'))
#Login to Graylog
uid_field = driver.find_element_by_name("username")
uid_field.clear()
uid_field.send_keys("admin")
password_field = driver.find_element_by_name("password")
password_field.clear()
password_field.send_keys("admin")
password_field.send_keys(Keys.RETURN)
WebDriverWait(driver, 30).until(expected_conditions.title_contains('Getting started'))
#Run tests
yield driver
finally:
driver.quit()
| import pytest
from selenium import webdriver
from selenium.webdriver.chrome.options import Options
from selenium.webdriver.common.keys import Keys
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.support import expected_conditions
import time
from webdriver_manager.chrome import ChromeDriverManager
@pytest.fixture(scope="session")
def chromedriver():
try:
options = Options()
options.headless = True
options.add_argument('--no-sandbox')
options.add_argument('--disable-dev-shm-usage')
options.add_argument("--disable-gpu")
driver = webdriver.Chrome(ChromeDriverManager().install(), options=options)
url = 'http://localhost:9000'
driver.get(url + "/gettingstarted")
WebDriverWait(driver, 30).until(expected_conditions.title_contains('Sign in'))
#Login to Graylog
uid_field = driver.find_element_by_name("username")
uid_field.clear()
uid_field.send_keys("admin")
password_field = driver.find_element_by_name("password")
password_field.clear()
password_field.send_keys("admin")
password_field.send_keys(Keys.RETURN)
WebDriverWait(driver, 30).until(expected_conditions.title_contains('Getting started'))
#Run tests
yield driver
finally:
driver.quit()
| Switch UI tests back to google chrome. | Switch UI tests back to google chrome.
| Python | apache-2.0 | Graylog2/graylog-ansible-role | import pytest
from selenium import webdriver
from selenium.webdriver.chrome.options import Options
from selenium.webdriver.common.keys import Keys
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.support import expected_conditions
import time
from webdriver_manager.chrome import ChromeDriverManager
- from webdriver_manager.utils import ChromeType
@pytest.fixture(scope="session")
def chromedriver():
try:
options = Options()
options.headless = True
options.add_argument('--no-sandbox')
options.add_argument('--disable-dev-shm-usage')
options.add_argument("--disable-gpu")
- driver = webdriver.Chrome(ChromeDriverManager(chrome_type=ChromeType.CHROMIUM).install(), options=options)
+ driver = webdriver.Chrome(ChromeDriverManager().install(), options=options)
url = 'http://localhost:9000'
driver.get(url + "/gettingstarted")
WebDriverWait(driver, 30).until(expected_conditions.title_contains('Sign in'))
#Login to Graylog
uid_field = driver.find_element_by_name("username")
uid_field.clear()
uid_field.send_keys("admin")
password_field = driver.find_element_by_name("password")
password_field.clear()
password_field.send_keys("admin")
password_field.send_keys(Keys.RETURN)
WebDriverWait(driver, 30).until(expected_conditions.title_contains('Getting started'))
#Run tests
yield driver
finally:
driver.quit()
| Switch UI tests back to google chrome. | ## Code Before:
import pytest
from selenium import webdriver
from selenium.webdriver.chrome.options import Options
from selenium.webdriver.common.keys import Keys
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.support import expected_conditions
import time
from webdriver_manager.chrome import ChromeDriverManager
from webdriver_manager.utils import ChromeType
@pytest.fixture(scope="session")
def chromedriver():
try:
options = Options()
options.headless = True
options.add_argument('--no-sandbox')
options.add_argument('--disable-dev-shm-usage')
options.add_argument("--disable-gpu")
driver = webdriver.Chrome(ChromeDriverManager(chrome_type=ChromeType.CHROMIUM).install(), options=options)
url = 'http://localhost:9000'
driver.get(url + "/gettingstarted")
WebDriverWait(driver, 30).until(expected_conditions.title_contains('Sign in'))
#Login to Graylog
uid_field = driver.find_element_by_name("username")
uid_field.clear()
uid_field.send_keys("admin")
password_field = driver.find_element_by_name("password")
password_field.clear()
password_field.send_keys("admin")
password_field.send_keys(Keys.RETURN)
WebDriverWait(driver, 30).until(expected_conditions.title_contains('Getting started'))
#Run tests
yield driver
finally:
driver.quit()
## Instruction:
Switch UI tests back to google chrome.
## Code After:
import pytest
from selenium import webdriver
from selenium.webdriver.chrome.options import Options
from selenium.webdriver.common.keys import Keys
from selenium.webdriver.support.ui import WebDriverWait
from selenium.webdriver.support import expected_conditions
import time
from webdriver_manager.chrome import ChromeDriverManager
@pytest.fixture(scope="session")
def chromedriver():
try:
options = Options()
options.headless = True
options.add_argument('--no-sandbox')
options.add_argument('--disable-dev-shm-usage')
options.add_argument("--disable-gpu")
driver = webdriver.Chrome(ChromeDriverManager().install(), options=options)
url = 'http://localhost:9000'
driver.get(url + "/gettingstarted")
WebDriverWait(driver, 30).until(expected_conditions.title_contains('Sign in'))
#Login to Graylog
uid_field = driver.find_element_by_name("username")
uid_field.clear()
uid_field.send_keys("admin")
password_field = driver.find_element_by_name("password")
password_field.clear()
password_field.send_keys("admin")
password_field.send_keys(Keys.RETURN)
WebDriverWait(driver, 30).until(expected_conditions.title_contains('Getting started'))
#Run tests
yield driver
finally:
driver.quit()
|
760506e88d22d86be818017fb6075abe7af2a068 | dactyl.py | dactyl.py | from slackbot.bot import respond_to
from slackbot.bot import listen_to
import re
import urllib
| from slackbot.bot import respond_to
from slackbot.bot import listen_to
import re
import urllib
def url_validator(url):
try:
code = urllib.urlopen(url).getcode()
if code == 200:
return True
except:
return False
def test_url(message, url):
if url_validator(url[1:len(url)-1]):
message.reply('VALID URL')
else:
message.reply('NOT VALID URL')
| Add url_validator function and respond aciton to test url | Add url_validator function and respond aciton to test url
| Python | mit | KrzysztofSendor/dactyl | from slackbot.bot import respond_to
from slackbot.bot import listen_to
import re
import urllib
+
+ def url_validator(url):
+ try:
+ code = urllib.urlopen(url).getcode()
+ if code == 200:
+ return True
+ except:
+ return False
+
+
+ def test_url(message, url):
+ if url_validator(url[1:len(url)-1]):
+ message.reply('VALID URL')
+ else:
+ message.reply('NOT VALID URL')
+ | Add url_validator function and respond aciton to test url | ## Code Before:
from slackbot.bot import respond_to
from slackbot.bot import listen_to
import re
import urllib
## Instruction:
Add url_validator function and respond aciton to test url
## Code After:
from slackbot.bot import respond_to
from slackbot.bot import listen_to
import re
import urllib
def url_validator(url):
try:
code = urllib.urlopen(url).getcode()
if code == 200:
return True
except:
return False
def test_url(message, url):
if url_validator(url[1:len(url)-1]):
message.reply('VALID URL')
else:
message.reply('NOT VALID URL')
|
1b8efb09ac512622ea3541d950ffc67b0a183178 | survey/signals.py | survey/signals.py | import django.dispatch
survey_completed = django.dispatch.Signal(providing_args=["instance", "data"])
| import django.dispatch
# providing_args=["instance", "data"]
survey_completed = django.dispatch.Signal()
| Remove puyrely documental providing-args argument | Remove puyrely documental providing-args argument
See https://docs.djangoproject.com/en/4.0/releases/3.1/#id2
| Python | agpl-3.0 | Pierre-Sassoulas/django-survey,Pierre-Sassoulas/django-survey,Pierre-Sassoulas/django-survey | import django.dispatch
- survey_completed = django.dispatch.Signal(providing_args=["instance", "data"])
+ # providing_args=["instance", "data"]
+ survey_completed = django.dispatch.Signal()
| Remove puyrely documental providing-args argument | ## Code Before:
import django.dispatch
survey_completed = django.dispatch.Signal(providing_args=["instance", "data"])
## Instruction:
Remove puyrely documental providing-args argument
## Code After:
import django.dispatch
# providing_args=["instance", "data"]
survey_completed = django.dispatch.Signal()
|
231657e2bbc81b8299cc91fd24dcd7394f74b4ec | python/dpu_utils/codeutils/identifiersplitting.py | python/dpu_utils/codeutils/identifiersplitting.py | from functools import lru_cache
from typing import List
import sys
REGEX_TEXT = ("(?<=[a-z0-9])(?=[A-Z])|"
"(?<=[A-Z0-9])(?=[A-Z][a-z])|"
"(?<=[0-9])(?=[a-zA-Z])|"
"(?<=[A-Za-z])(?=[0-9])|"
"(?<=[@$])(?=[a-zA-Z0-9])|"
"(?<=[a-zA-Z0-9])(?=[@$])|"
"_")
if sys.version_info >= (3, 7):
import re
SPLIT_REGEX = re.compile(REGEX_TEXT)
else:
import regex
SPLIT_REGEX = regex.compile("(?V1)"+REGEX_TEXT)
@lru_cache(maxsize=5000)
def split_identifier_into_parts(identifier: str) -> List[str]:
"""
Split a single identifier into parts on snake_case and camelCase
"""
identifier_parts = list(s for s in SPLIT_REGEX.split(identifier) if len(s)>0)
if len(identifier_parts) == 0:
return [identifier]
return identifier_parts
| from functools import lru_cache
from typing import List
import sys
REGEX_TEXT = ("(?<=[a-z0-9])(?=[A-Z])|"
"(?<=[A-Z0-9])(?=[A-Z][a-z])|"
"(?<=[0-9])(?=[a-zA-Z])|"
"(?<=[A-Za-z])(?=[0-9])|"
"(?<=[@$.'\"])(?=[a-zA-Z0-9])|"
"(?<=[a-zA-Z0-9])(?=[@$.'\"])|"
"_|\\s+")
if sys.version_info >= (3, 7):
import re
SPLIT_REGEX = re.compile(REGEX_TEXT)
else:
import regex
SPLIT_REGEX = regex.compile("(?V1)"+REGEX_TEXT)
@lru_cache(maxsize=5000)
def split_identifier_into_parts(identifier: str) -> List[str]:
"""
Split a single identifier into parts on snake_case and camelCase
"""
identifier_parts = list(s.lower() for s in SPLIT_REGEX.split(identifier) if len(s)>0)
if len(identifier_parts) == 0:
return [identifier]
return identifier_parts
| Revert to some of the previous behavior for characters that shouldn't appear in identifiers. | Revert to some of the previous behavior for characters that shouldn't appear in identifiers.
| Python | mit | microsoft/dpu-utils,microsoft/dpu-utils | from functools import lru_cache
from typing import List
import sys
REGEX_TEXT = ("(?<=[a-z0-9])(?=[A-Z])|"
"(?<=[A-Z0-9])(?=[A-Z][a-z])|"
"(?<=[0-9])(?=[a-zA-Z])|"
"(?<=[A-Za-z])(?=[0-9])|"
- "(?<=[@$])(?=[a-zA-Z0-9])|"
+ "(?<=[@$.'\"])(?=[a-zA-Z0-9])|"
- "(?<=[a-zA-Z0-9])(?=[@$])|"
+ "(?<=[a-zA-Z0-9])(?=[@$.'\"])|"
- "_")
+ "_|\\s+")
if sys.version_info >= (3, 7):
import re
SPLIT_REGEX = re.compile(REGEX_TEXT)
else:
import regex
SPLIT_REGEX = regex.compile("(?V1)"+REGEX_TEXT)
@lru_cache(maxsize=5000)
def split_identifier_into_parts(identifier: str) -> List[str]:
"""
Split a single identifier into parts on snake_case and camelCase
"""
- identifier_parts = list(s for s in SPLIT_REGEX.split(identifier) if len(s)>0)
+ identifier_parts = list(s.lower() for s in SPLIT_REGEX.split(identifier) if len(s)>0)
if len(identifier_parts) == 0:
return [identifier]
return identifier_parts
| Revert to some of the previous behavior for characters that shouldn't appear in identifiers. | ## Code Before:
from functools import lru_cache
from typing import List
import sys
REGEX_TEXT = ("(?<=[a-z0-9])(?=[A-Z])|"
"(?<=[A-Z0-9])(?=[A-Z][a-z])|"
"(?<=[0-9])(?=[a-zA-Z])|"
"(?<=[A-Za-z])(?=[0-9])|"
"(?<=[@$])(?=[a-zA-Z0-9])|"
"(?<=[a-zA-Z0-9])(?=[@$])|"
"_")
if sys.version_info >= (3, 7):
import re
SPLIT_REGEX = re.compile(REGEX_TEXT)
else:
import regex
SPLIT_REGEX = regex.compile("(?V1)"+REGEX_TEXT)
@lru_cache(maxsize=5000)
def split_identifier_into_parts(identifier: str) -> List[str]:
"""
Split a single identifier into parts on snake_case and camelCase
"""
identifier_parts = list(s for s in SPLIT_REGEX.split(identifier) if len(s)>0)
if len(identifier_parts) == 0:
return [identifier]
return identifier_parts
## Instruction:
Revert to some of the previous behavior for characters that shouldn't appear in identifiers.
## Code After:
from functools import lru_cache
from typing import List
import sys
REGEX_TEXT = ("(?<=[a-z0-9])(?=[A-Z])|"
"(?<=[A-Z0-9])(?=[A-Z][a-z])|"
"(?<=[0-9])(?=[a-zA-Z])|"
"(?<=[A-Za-z])(?=[0-9])|"
"(?<=[@$.'\"])(?=[a-zA-Z0-9])|"
"(?<=[a-zA-Z0-9])(?=[@$.'\"])|"
"_|\\s+")
if sys.version_info >= (3, 7):
import re
SPLIT_REGEX = re.compile(REGEX_TEXT)
else:
import regex
SPLIT_REGEX = regex.compile("(?V1)"+REGEX_TEXT)
@lru_cache(maxsize=5000)
def split_identifier_into_parts(identifier: str) -> List[str]:
"""
Split a single identifier into parts on snake_case and camelCase
"""
identifier_parts = list(s.lower() for s in SPLIT_REGEX.split(identifier) if len(s)>0)
if len(identifier_parts) == 0:
return [identifier]
return identifier_parts
|
81460f88ee19fb736dfc3453df2905f0ba4b3974 | common/permissions.py | common/permissions.py | from rest_framework.permissions import BasePermission
class ObjectHasTokenUser(BasePermission):
"""
The object's user matches the token's user.
"""
def has_object_permission(self, request, view, obj):
token = request.auth
if not token:
return False
if not hasattr(token, 'scope'):
assert False, ('TokenHasReadWriteScope requires the'
'`OAuth2Authentication` authentication '
'class to be used.')
if hasattr(obj, 'user'):
print 'token.user', token.user
print 'obj.user', obj.user
return token.user == obj.user
| from rest_framework.permissions import BasePermission
class ObjectHasTokenUser(BasePermission):
"""
The object's user matches the token's user.
"""
def has_object_permission(self, request, view, obj):
token = request.auth
if not token:
return False
if not hasattr(token, 'scope'):
assert False, ('ObjectHasTokenUser requires the'
'`OAuth2Authentication` authentication '
'class to be used.')
if hasattr(obj, 'user'):
return token.user == obj.user
| Remove debugging code, fix typo | Remove debugging code, fix typo
| Python | mit | PersonalGenomesOrg/open-humans,PersonalGenomesOrg/open-humans,PersonalGenomesOrg/open-humans,OpenHumans/open-humans,OpenHumans/open-humans,OpenHumans/open-humans,PersonalGenomesOrg/open-humans,OpenHumans/open-humans | from rest_framework.permissions import BasePermission
class ObjectHasTokenUser(BasePermission):
"""
The object's user matches the token's user.
"""
def has_object_permission(self, request, view, obj):
token = request.auth
if not token:
return False
if not hasattr(token, 'scope'):
- assert False, ('TokenHasReadWriteScope requires the'
+ assert False, ('ObjectHasTokenUser requires the'
'`OAuth2Authentication` authentication '
'class to be used.')
if hasattr(obj, 'user'):
- print 'token.user', token.user
- print 'obj.user', obj.user
-
return token.user == obj.user
| Remove debugging code, fix typo | ## Code Before:
from rest_framework.permissions import BasePermission
class ObjectHasTokenUser(BasePermission):
"""
The object's user matches the token's user.
"""
def has_object_permission(self, request, view, obj):
token = request.auth
if not token:
return False
if not hasattr(token, 'scope'):
assert False, ('TokenHasReadWriteScope requires the'
'`OAuth2Authentication` authentication '
'class to be used.')
if hasattr(obj, 'user'):
print 'token.user', token.user
print 'obj.user', obj.user
return token.user == obj.user
## Instruction:
Remove debugging code, fix typo
## Code After:
from rest_framework.permissions import BasePermission
class ObjectHasTokenUser(BasePermission):
"""
The object's user matches the token's user.
"""
def has_object_permission(self, request, view, obj):
token = request.auth
if not token:
return False
if not hasattr(token, 'scope'):
assert False, ('ObjectHasTokenUser requires the'
'`OAuth2Authentication` authentication '
'class to be used.')
if hasattr(obj, 'user'):
return token.user == obj.user
|
65336509829a42b91b000d2e423ed4581ac61c98 | app/mpv.py | app/mpv.py | import sys
import json
import struct
import subprocess
# Read a message from stdin and decode it.
def getMessage():
rawLength = sys.stdin.read(4)
if len(rawLength) == 0:
sys.exit(0)
messageLength = struct.unpack('@I', rawLength)[0]
message = sys.stdin.read(messageLength)
return json.loads(message)
# Encode a message for transmission,
# given its content.
def encodeMessage(messageContent):
encodedContent = json.dumps(messageContent)
encodedLength = struct.pack('@I', len(encodedContent))
return {'length': encodedLength, 'content': encodedContent}
# Send an encoded message to stdout
def sendMessage(encodedMessage):
sys.stdout.write(encodedMessage['length'])
sys.stdout.write(encodedMessage['content'])
sys.stdout.flush()
while True:
mpv_args = getMessage()
if (len(mpv_args) > 1):
subprocess.call(["mpv", mpv_args])
|
import sys
import json
import struct
import subprocess
import shlex
# Read a message from stdin and decode it.
def getMessage():
rawLength = sys.stdin.read(4)
if len(rawLength) == 0:
sys.exit(0)
messageLength = struct.unpack('@I', rawLength)[0]
message = sys.stdin.read(messageLength)
return json.loads(message)
# Encode a message for transmission,
# given its content.
def encodeMessage(messageContent):
encodedContent = json.dumps(messageContent)
encodedLength = struct.pack('@I', len(encodedContent))
return {'length': encodedLength, 'content': encodedContent}
# Send an encoded message to stdout
def sendMessage(encodedMessage):
sys.stdout.write(encodedMessage['length'])
sys.stdout.write(encodedMessage['content'])
sys.stdout.flush()
while True:
mpv_args = getMessage()
if (len(mpv_args) > 1):
args = shlex.split("mpv " + mpv_args)
subprocess.call(args)
sys.exit(0)
| Handle shell args in python scripts | Handle shell args in python scripts
| Python | mit | vayan/external-video,vayan/external-video,vayan/external-video,vayan/external-video | +
import sys
import json
import struct
import subprocess
+ import shlex
# Read a message from stdin and decode it.
def getMessage():
rawLength = sys.stdin.read(4)
if len(rawLength) == 0:
sys.exit(0)
- messageLength = struct.unpack('@I', rawLength)[0]
+ messageLength = struct.unpack('@I', rawLength)[0]
- message = sys.stdin.read(messageLength)
+ message = sys.stdin.read(messageLength)
- return json.loads(message)
+ return json.loads(message)
# Encode a message for transmission,
# given its content.
def encodeMessage(messageContent):
encodedContent = json.dumps(messageContent)
encodedLength = struct.pack('@I', len(encodedContent))
return {'length': encodedLength, 'content': encodedContent}
# Send an encoded message to stdout
def sendMessage(encodedMessage):
sys.stdout.write(encodedMessage['length'])
sys.stdout.write(encodedMessage['content'])
sys.stdout.flush()
while True:
mpv_args = getMessage()
if (len(mpv_args) > 1):
+ args = shlex.split("mpv " + mpv_args)
- subprocess.call(["mpv", mpv_args])
+ subprocess.call(args)
+ sys.exit(0)
| Handle shell args in python scripts | ## Code Before:
import sys
import json
import struct
import subprocess
# Read a message from stdin and decode it.
def getMessage():
rawLength = sys.stdin.read(4)
if len(rawLength) == 0:
sys.exit(0)
messageLength = struct.unpack('@I', rawLength)[0]
message = sys.stdin.read(messageLength)
return json.loads(message)
# Encode a message for transmission,
# given its content.
def encodeMessage(messageContent):
encodedContent = json.dumps(messageContent)
encodedLength = struct.pack('@I', len(encodedContent))
return {'length': encodedLength, 'content': encodedContent}
# Send an encoded message to stdout
def sendMessage(encodedMessage):
sys.stdout.write(encodedMessage['length'])
sys.stdout.write(encodedMessage['content'])
sys.stdout.flush()
while True:
mpv_args = getMessage()
if (len(mpv_args) > 1):
subprocess.call(["mpv", mpv_args])
## Instruction:
Handle shell args in python scripts
## Code After:
import sys
import json
import struct
import subprocess
import shlex
# Read a message from stdin and decode it.
def getMessage():
rawLength = sys.stdin.read(4)
if len(rawLength) == 0:
sys.exit(0)
messageLength = struct.unpack('@I', rawLength)[0]
message = sys.stdin.read(messageLength)
return json.loads(message)
# Encode a message for transmission,
# given its content.
def encodeMessage(messageContent):
encodedContent = json.dumps(messageContent)
encodedLength = struct.pack('@I', len(encodedContent))
return {'length': encodedLength, 'content': encodedContent}
# Send an encoded message to stdout
def sendMessage(encodedMessage):
sys.stdout.write(encodedMessage['length'])
sys.stdout.write(encodedMessage['content'])
sys.stdout.flush()
while True:
mpv_args = getMessage()
if (len(mpv_args) > 1):
args = shlex.split("mpv " + mpv_args)
subprocess.call(args)
sys.exit(0)
|
83d5cc3b4ffb4759e8e073d04299a55802df09a8 | src/ansible/views.py | src/ansible/views.py | from django.shortcuts import get_object_or_404, render
from django.http import HttpResponseRedirect, HttpResponse
from .models import Playbook
def index(request):
return "200"
| from django.shortcuts import get_object_or_404, render
from django.http import HttpResponseRedirect, HttpResponse
from .models import Playbook
def index(request):
return HttpResponse("200")
| Fix return to use HttpResponse | Fix return to use HttpResponse
| Python | bsd-3-clause | lozadaOmr/ansible-admin,lozadaOmr/ansible-admin,lozadaOmr/ansible-admin | from django.shortcuts import get_object_or_404, render
from django.http import HttpResponseRedirect, HttpResponse
from .models import Playbook
def index(request):
- return "200"
+ return HttpResponse("200")
| Fix return to use HttpResponse | ## Code Before:
from django.shortcuts import get_object_or_404, render
from django.http import HttpResponseRedirect, HttpResponse
from .models import Playbook
def index(request):
return "200"
## Instruction:
Fix return to use HttpResponse
## Code After:
from django.shortcuts import get_object_or_404, render
from django.http import HttpResponseRedirect, HttpResponse
from .models import Playbook
def index(request):
return HttpResponse("200")
|
88e5ecad9966057203a9cbecaeaecdca3e76b6da | tests/fake_filesystem.py | tests/fake_filesystem.py | import os
import stat
from StringIO import StringIO
from types import StringTypes
import paramiko as ssh
class FakeFile(StringIO):
def __init__(self, value=None, path=None):
init = lambda x: StringIO.__init__(self, x)
if value is None:
init("")
ftype = 'dir'
size = 4096
else:
init(value)
ftype = 'file'
size = len(value)
attr = ssh.SFTPAttributes()
attr.st_mode = {'file': stat.S_IFREG, 'dir': stat.S_IFDIR}[ftype]
attr.st_size = size
attr.filename = os.path.basename(path)
self.attributes = attr
def __str__(self):
return self.getvalue()
def write(self, value):
StringIO.write(self, value)
self.attributes.st_size = len(self.getvalue())
class FakeFilesystem(dict):
def __init__(self, d=None):
# Replicate input dictionary using our custom __setitem__
d = d or {}
for key, value in d.iteritems():
self[key] = value
def __setitem__(self, key, value):
if isinstance(value, StringTypes) or value is None:
value = FakeFile(value, key)
super(FakeFilesystem, self).__setitem__(key, value)
| import os
import stat
from StringIO import StringIO
from types import StringTypes
import paramiko as ssh
class FakeFile(StringIO):
def __init__(self, value=None, path=None):
init = lambda x: StringIO.__init__(self, x)
if value is None:
init("")
ftype = 'dir'
size = 4096
else:
init(value)
ftype = 'file'
size = len(value)
attr = ssh.SFTPAttributes()
attr.st_mode = {'file': stat.S_IFREG, 'dir': stat.S_IFDIR}[ftype]
attr.st_size = size
attr.filename = os.path.basename(path)
self.attributes = attr
def __str__(self):
return self.getvalue()
def write(self, value):
StringIO.write(self, value)
self.attributes.st_size = len(self.getvalue())
def close(self):
"""
Always hold fake files open.
"""
pass
class FakeFilesystem(dict):
def __init__(self, d=None):
# Replicate input dictionary using our custom __setitem__
d = d or {}
for key, value in d.iteritems():
self[key] = value
def __setitem__(self, key, value):
if isinstance(value, StringTypes) or value is None:
value = FakeFile(value, key)
super(FakeFilesystem, self).__setitem__(key, value)
| Define noop close() for FakeFile | Define noop close() for FakeFile
| Python | bsd-2-clause | kxxoling/fabric,rodrigc/fabric,qinrong/fabric,elijah513/fabric,bspink/fabric,MjAbuz/fabric,cmattoon/fabric,hrubi/fabric,felix-d/fabric,askulkarni2/fabric,SamuelMarks/fabric,mathiasertl/fabric,tekapo/fabric,StackStorm/fabric,ploxiln/fabric,kmonsoor/fabric,raimon49/fabric,haridsv/fabric,bitprophet/fabric,fernandezcuesta/fabric,itoed/fabric,rane-hs/fabric-py3,sdelements/fabric,likesxuqiang/fabric,bitmonk/fabric,getsentry/fabric,opavader/fabric,jaraco/fabric,xLegoz/fabric,TarasRudnyk/fabric,pgroudas/fabric,akaariai/fabric,rbramwell/fabric,amaniak/fabric,cgvarela/fabric,tolbkni/fabric,pashinin/fabric | import os
import stat
from StringIO import StringIO
from types import StringTypes
import paramiko as ssh
class FakeFile(StringIO):
def __init__(self, value=None, path=None):
init = lambda x: StringIO.__init__(self, x)
if value is None:
init("")
ftype = 'dir'
size = 4096
else:
init(value)
ftype = 'file'
size = len(value)
attr = ssh.SFTPAttributes()
attr.st_mode = {'file': stat.S_IFREG, 'dir': stat.S_IFDIR}[ftype]
attr.st_size = size
attr.filename = os.path.basename(path)
self.attributes = attr
def __str__(self):
return self.getvalue()
def write(self, value):
StringIO.write(self, value)
self.attributes.st_size = len(self.getvalue())
+ def close(self):
+ """
+ Always hold fake files open.
+ """
+ pass
+
class FakeFilesystem(dict):
def __init__(self, d=None):
# Replicate input dictionary using our custom __setitem__
d = d or {}
for key, value in d.iteritems():
self[key] = value
def __setitem__(self, key, value):
if isinstance(value, StringTypes) or value is None:
value = FakeFile(value, key)
super(FakeFilesystem, self).__setitem__(key, value)
| Define noop close() for FakeFile | ## Code Before:
import os
import stat
from StringIO import StringIO
from types import StringTypes
import paramiko as ssh
class FakeFile(StringIO):
def __init__(self, value=None, path=None):
init = lambda x: StringIO.__init__(self, x)
if value is None:
init("")
ftype = 'dir'
size = 4096
else:
init(value)
ftype = 'file'
size = len(value)
attr = ssh.SFTPAttributes()
attr.st_mode = {'file': stat.S_IFREG, 'dir': stat.S_IFDIR}[ftype]
attr.st_size = size
attr.filename = os.path.basename(path)
self.attributes = attr
def __str__(self):
return self.getvalue()
def write(self, value):
StringIO.write(self, value)
self.attributes.st_size = len(self.getvalue())
class FakeFilesystem(dict):
def __init__(self, d=None):
# Replicate input dictionary using our custom __setitem__
d = d or {}
for key, value in d.iteritems():
self[key] = value
def __setitem__(self, key, value):
if isinstance(value, StringTypes) or value is None:
value = FakeFile(value, key)
super(FakeFilesystem, self).__setitem__(key, value)
## Instruction:
Define noop close() for FakeFile
## Code After:
import os
import stat
from StringIO import StringIO
from types import StringTypes
import paramiko as ssh
class FakeFile(StringIO):
def __init__(self, value=None, path=None):
init = lambda x: StringIO.__init__(self, x)
if value is None:
init("")
ftype = 'dir'
size = 4096
else:
init(value)
ftype = 'file'
size = len(value)
attr = ssh.SFTPAttributes()
attr.st_mode = {'file': stat.S_IFREG, 'dir': stat.S_IFDIR}[ftype]
attr.st_size = size
attr.filename = os.path.basename(path)
self.attributes = attr
def __str__(self):
return self.getvalue()
def write(self, value):
StringIO.write(self, value)
self.attributes.st_size = len(self.getvalue())
def close(self):
"""
Always hold fake files open.
"""
pass
class FakeFilesystem(dict):
def __init__(self, d=None):
# Replicate input dictionary using our custom __setitem__
d = d or {}
for key, value in d.iteritems():
self[key] = value
def __setitem__(self, key, value):
if isinstance(value, StringTypes) or value is None:
value = FakeFile(value, key)
super(FakeFilesystem, self).__setitem__(key, value)
|
f45b3e73b6258c99aed2bff2e7350f1c797ff849 | providers/provider.py | providers/provider.py | import copy
import json
import requests
import html5lib
from application import APPLICATION as APP
# Be compatible with python 2 and 3
try:
from urllib import urlencode
except ImportError:
from urllib.parse import urlencode
class BaseProvider(object):
# ==== HELPER METHODS ====
def parse_html(self, url, css_selector, timeout=60, cache=True):
html = self._http_get(url, timeout=timeout, cache=cache)
document = html5lib.parse(html)
results = document.cssselect(css_selector)
data = [result.text_content() for result in results]
return data
def traverse_json(self, data, path):
if not path:
return data
new_data = copy.copy(data)
for item in path.split("."):
if item.isdigit():
item = int(item)
try:
new_data = new_data[item]
except (IndexError, KeyError):
return {}
return new_data
def parse_json(self, url, path=None, timeout=60, cache=True):
data = self._http_get(url, timeout=timeout, cache=cache)
data = json.loads(data)
data = self.traverse_json(data, path)
return data
def urlencode(self, data):
return urlencode(data)
# ==== PRIVATE METHODS ====
def _http_get(self, url, timeout=60, cache=True):
base = requests if not cache else APP.setting("WEBCACHE")
response = base.get(url, timeout=timeout)
return response.text
| import copy
import json
from urllib.parse import urlencode
import html5lib
import requests
from application import APPLICATION as APP
class BaseProvider(object):
# ==== HELPER METHODS ====
def parse_html(self, url, css_selector, timeout=60, cache=True):
html = self._http_get(url, timeout=timeout, cache=cache)
document = html5lib.parse(html)
results = document.cssselect(css_selector)
data = [result.text_content() for result in results]
return data
def traverse_json(self, data, path):
if not path:
return data
new_data = copy.copy(data)
for item in path.split("."):
if item.isdigit():
item = int(item)
try:
new_data = new_data[item]
except (IndexError, KeyError):
return {}
return new_data
def parse_json(self, url, path=None, timeout=60, cache=True):
data = self._http_get(url, timeout=timeout, cache=cache)
data = json.loads(data)
data = self.traverse_json(data, path)
return data
def urlencode(self, data):
return urlencode(data)
# ==== PRIVATE METHODS ====
def _http_get(self, url, timeout=60, cache=True):
base = requests if not cache else APP.setting("WEBCACHE")
response = base.get(url, timeout=timeout)
return response.text
| Remove support for Python 2. | Remove support for Python 2.
| Python | mit | EmilStenstrom/nephele | import copy
import json
+ from urllib.parse import urlencode
+
+ import html5lib
import requests
- import html5lib
from application import APPLICATION as APP
- # Be compatible with python 2 and 3
- try:
- from urllib import urlencode
- except ImportError:
- from urllib.parse import urlencode
class BaseProvider(object):
# ==== HELPER METHODS ====
def parse_html(self, url, css_selector, timeout=60, cache=True):
html = self._http_get(url, timeout=timeout, cache=cache)
document = html5lib.parse(html)
results = document.cssselect(css_selector)
data = [result.text_content() for result in results]
return data
def traverse_json(self, data, path):
if not path:
return data
new_data = copy.copy(data)
for item in path.split("."):
if item.isdigit():
item = int(item)
try:
new_data = new_data[item]
except (IndexError, KeyError):
return {}
return new_data
def parse_json(self, url, path=None, timeout=60, cache=True):
data = self._http_get(url, timeout=timeout, cache=cache)
data = json.loads(data)
data = self.traverse_json(data, path)
return data
def urlencode(self, data):
return urlencode(data)
# ==== PRIVATE METHODS ====
def _http_get(self, url, timeout=60, cache=True):
base = requests if not cache else APP.setting("WEBCACHE")
response = base.get(url, timeout=timeout)
return response.text
| Remove support for Python 2. | ## Code Before:
import copy
import json
import requests
import html5lib
from application import APPLICATION as APP
# Be compatible with python 2 and 3
try:
from urllib import urlencode
except ImportError:
from urllib.parse import urlencode
class BaseProvider(object):
# ==== HELPER METHODS ====
def parse_html(self, url, css_selector, timeout=60, cache=True):
html = self._http_get(url, timeout=timeout, cache=cache)
document = html5lib.parse(html)
results = document.cssselect(css_selector)
data = [result.text_content() for result in results]
return data
def traverse_json(self, data, path):
if not path:
return data
new_data = copy.copy(data)
for item in path.split("."):
if item.isdigit():
item = int(item)
try:
new_data = new_data[item]
except (IndexError, KeyError):
return {}
return new_data
def parse_json(self, url, path=None, timeout=60, cache=True):
data = self._http_get(url, timeout=timeout, cache=cache)
data = json.loads(data)
data = self.traverse_json(data, path)
return data
def urlencode(self, data):
return urlencode(data)
# ==== PRIVATE METHODS ====
def _http_get(self, url, timeout=60, cache=True):
base = requests if not cache else APP.setting("WEBCACHE")
response = base.get(url, timeout=timeout)
return response.text
## Instruction:
Remove support for Python 2.
## Code After:
import copy
import json
from urllib.parse import urlencode
import html5lib
import requests
from application import APPLICATION as APP
class BaseProvider(object):
# ==== HELPER METHODS ====
def parse_html(self, url, css_selector, timeout=60, cache=True):
html = self._http_get(url, timeout=timeout, cache=cache)
document = html5lib.parse(html)
results = document.cssselect(css_selector)
data = [result.text_content() for result in results]
return data
def traverse_json(self, data, path):
if not path:
return data
new_data = copy.copy(data)
for item in path.split("."):
if item.isdigit():
item = int(item)
try:
new_data = new_data[item]
except (IndexError, KeyError):
return {}
return new_data
def parse_json(self, url, path=None, timeout=60, cache=True):
data = self._http_get(url, timeout=timeout, cache=cache)
data = json.loads(data)
data = self.traverse_json(data, path)
return data
def urlencode(self, data):
return urlencode(data)
# ==== PRIVATE METHODS ====
def _http_get(self, url, timeout=60, cache=True):
base = requests if not cache else APP.setting("WEBCACHE")
response = base.get(url, timeout=timeout)
return response.text
|
cdd32dd3e346f72f823cc5d3f59c79c027db65c8 | common.py | common.py | """Functions common to other modules."""
import json
import os
import re
import time
import urllib.request
from settings import net
def clean(name):
"""Strip all [^a-zA-Z0-9_] characters and convert to lowercase."""
return re.sub(r"\W", r"", name, flags=re.ASCII).lower()
def exists(path):
"""Check to see if a path exists."""
return True if os.path.exists(path) else False
def ls(path):
"""The contents of a directory."""
return os.listdir(path)
def mkdir(path):
"""Create the given directory path if it doesn't already exist."""
os.makedirs(path, exist_ok=True)
return path
def open_url(url, task):
"""Retrieve data from the specified url."""
for attempt in range(0, net.retries):
try:
return urllib.request.urlopen(url)
except OSError:
print("Error: {} (retry in {}s)".format(task, net.wait))
time.sleep(net.wait)
raise ConnectionError("Halted: Unable to access resource")
def urlopen_json(url, task):
"""Retrieve json data from the specified url."""
for attempt in range(0, net.retries):
try:
reply = urllib.request.urlopen(url)
reply = json.loads(reply.read().decode())
return reply["DATA"]["RECORD"]
except:
print("Error: {} (retry in {}s)".format(task, net.wait))
time.sleep(net.wait)
raise ConnectionError("Halted: Unable to access resource")
| """Functions common to other modules."""
import json
import os
import re
import time
import urllib.request
from settings import net
def clean(name):
"""Strip all [^a-zA-Z0-9_] characters and convert to lowercase."""
return re.sub(r"\W", r"", name, flags=re.ASCII).lower()
def exists(path):
"""Check to see if a path exists."""
return True if os.path.exists(path) else False
def ls(path):
"""The contents of a directory."""
return os.listdir(path)
def mkdir(path):
"""Create the given directory path if it doesn't already exist."""
os.makedirs(path, exist_ok=True)
return path
def open_url(url, task):
"""Retrieve data from the specified url."""
for attempt in range(0, net.retries):
try:
return urllib.request.urlopen(url)
except OSError:
print("Error: {} (retry in {}s)".format(task, net.wait))
time.sleep(net.wait)
raise ConnectionError("Halted: Unable to access resource")
def urlopen_json(url, task="Unknown task"):
"""Retrieve json data from the specified url."""
for attempt in range(0, net.retries):
try:
reply = urllib.request.urlopen(url)
reply = json.loads(reply.read().decode())
return reply["DATA"]["RECORD"]
except:
print("Error: {} (retry in {}s)".format(task, net.wait))
time.sleep(net.wait)
raise ConnectionError("Halted: Unable to access resource")
| Make task an optional argument. | Make task an optional argument.
| Python | bsd-2-clause | chingc/DJRivals,chingc/DJRivals | """Functions common to other modules."""
import json
import os
import re
import time
import urllib.request
from settings import net
def clean(name):
"""Strip all [^a-zA-Z0-9_] characters and convert to lowercase."""
return re.sub(r"\W", r"", name, flags=re.ASCII).lower()
def exists(path):
"""Check to see if a path exists."""
return True if os.path.exists(path) else False
def ls(path):
"""The contents of a directory."""
return os.listdir(path)
def mkdir(path):
"""Create the given directory path if it doesn't already exist."""
os.makedirs(path, exist_ok=True)
return path
def open_url(url, task):
"""Retrieve data from the specified url."""
for attempt in range(0, net.retries):
try:
return urllib.request.urlopen(url)
except OSError:
print("Error: {} (retry in {}s)".format(task, net.wait))
time.sleep(net.wait)
raise ConnectionError("Halted: Unable to access resource")
- def urlopen_json(url, task):
+ def urlopen_json(url, task="Unknown task"):
"""Retrieve json data from the specified url."""
for attempt in range(0, net.retries):
try:
reply = urllib.request.urlopen(url)
reply = json.loads(reply.read().decode())
return reply["DATA"]["RECORD"]
except:
print("Error: {} (retry in {}s)".format(task, net.wait))
time.sleep(net.wait)
raise ConnectionError("Halted: Unable to access resource")
| Make task an optional argument. | ## Code Before:
"""Functions common to other modules."""
import json
import os
import re
import time
import urllib.request
from settings import net
def clean(name):
"""Strip all [^a-zA-Z0-9_] characters and convert to lowercase."""
return re.sub(r"\W", r"", name, flags=re.ASCII).lower()
def exists(path):
"""Check to see if a path exists."""
return True if os.path.exists(path) else False
def ls(path):
"""The contents of a directory."""
return os.listdir(path)
def mkdir(path):
"""Create the given directory path if it doesn't already exist."""
os.makedirs(path, exist_ok=True)
return path
def open_url(url, task):
"""Retrieve data from the specified url."""
for attempt in range(0, net.retries):
try:
return urllib.request.urlopen(url)
except OSError:
print("Error: {} (retry in {}s)".format(task, net.wait))
time.sleep(net.wait)
raise ConnectionError("Halted: Unable to access resource")
def urlopen_json(url, task):
"""Retrieve json data from the specified url."""
for attempt in range(0, net.retries):
try:
reply = urllib.request.urlopen(url)
reply = json.loads(reply.read().decode())
return reply["DATA"]["RECORD"]
except:
print("Error: {} (retry in {}s)".format(task, net.wait))
time.sleep(net.wait)
raise ConnectionError("Halted: Unable to access resource")
## Instruction:
Make task an optional argument.
## Code After:
"""Functions common to other modules."""
import json
import os
import re
import time
import urllib.request
from settings import net
def clean(name):
"""Strip all [^a-zA-Z0-9_] characters and convert to lowercase."""
return re.sub(r"\W", r"", name, flags=re.ASCII).lower()
def exists(path):
"""Check to see if a path exists."""
return True if os.path.exists(path) else False
def ls(path):
"""The contents of a directory."""
return os.listdir(path)
def mkdir(path):
"""Create the given directory path if it doesn't already exist."""
os.makedirs(path, exist_ok=True)
return path
def open_url(url, task):
"""Retrieve data from the specified url."""
for attempt in range(0, net.retries):
try:
return urllib.request.urlopen(url)
except OSError:
print("Error: {} (retry in {}s)".format(task, net.wait))
time.sleep(net.wait)
raise ConnectionError("Halted: Unable to access resource")
def urlopen_json(url, task="Unknown task"):
"""Retrieve json data from the specified url."""
for attempt in range(0, net.retries):
try:
reply = urllib.request.urlopen(url)
reply = json.loads(reply.read().decode())
return reply["DATA"]["RECORD"]
except:
print("Error: {} (retry in {}s)".format(task, net.wait))
time.sleep(net.wait)
raise ConnectionError("Halted: Unable to access resource")
|
150e338b7d2793c434d7e2f21aef061f35634476 | openspending/test/__init__.py | openspending/test/__init__.py |
import os
import sys
from paste.deploy import appconfig
from openspending import mongo
from helpers import clean_all
__all__ = ['TestCase', 'DatabaseTestCase']
here_dir = os.getcwd()
config = appconfig('config:test.ini', relative_to=here_dir)
mongo.configure(config)
class TestCase(object):
def setup(self):
pass
def teardown(self):
pass
class DatabaseTestCase(TestCase):
def teardown(self):
clean_all()
super(DatabaseTestCase, self).teardown() |
import os
import sys
from pylons import config
from openspending import mongo
from .helpers import clean_all
__all__ = ['TestCase', 'DatabaseTestCase']
mongo.configure(config)
class TestCase(object):
def setup(self):
pass
def teardown(self):
pass
class DatabaseTestCase(TestCase):
def teardown(self):
clean_all()
super(DatabaseTestCase, self).teardown() | Use config given on command line | Use config given on command line
| Python | agpl-3.0 | pudo/spendb,openspending/spendb,openspending/spendb,nathanhilbert/FPA_Core,pudo/spendb,johnjohndoe/spendb,CivicVision/datahub,USStateDept/FPA_Core,johnjohndoe/spendb,USStateDept/FPA_Core,CivicVision/datahub,nathanhilbert/FPA_Core,johnjohndoe/spendb,openspending/spendb,USStateDept/FPA_Core,spendb/spendb,nathanhilbert/FPA_Core,spendb/spendb,pudo/spendb,spendb/spendb,CivicVision/datahub |
import os
import sys
- from paste.deploy import appconfig
+ from pylons import config
from openspending import mongo
- from helpers import clean_all
+ from .helpers import clean_all
__all__ = ['TestCase', 'DatabaseTestCase']
- here_dir = os.getcwd()
- config = appconfig('config:test.ini', relative_to=here_dir)
mongo.configure(config)
class TestCase(object):
def setup(self):
pass
def teardown(self):
pass
class DatabaseTestCase(TestCase):
def teardown(self):
clean_all()
super(DatabaseTestCase, self).teardown() | Use config given on command line | ## Code Before:
import os
import sys
from paste.deploy import appconfig
from openspending import mongo
from helpers import clean_all
__all__ = ['TestCase', 'DatabaseTestCase']
here_dir = os.getcwd()
config = appconfig('config:test.ini', relative_to=here_dir)
mongo.configure(config)
class TestCase(object):
def setup(self):
pass
def teardown(self):
pass
class DatabaseTestCase(TestCase):
def teardown(self):
clean_all()
super(DatabaseTestCase, self).teardown()
## Instruction:
Use config given on command line
## Code After:
import os
import sys
from pylons import config
from openspending import mongo
from .helpers import clean_all
__all__ = ['TestCase', 'DatabaseTestCase']
mongo.configure(config)
class TestCase(object):
def setup(self):
pass
def teardown(self):
pass
class DatabaseTestCase(TestCase):
def teardown(self):
clean_all()
super(DatabaseTestCase, self).teardown() |
37a0cb41a88114ab9edb514e29447756b0c3e92a | tests/test_cli.py | tests/test_cli.py |
from click.testing import CliRunner
import pytest
from cibopath.cli import main
from cibopath import __version__
runner = CliRunner()
@pytest.fixture(params=['-V', '--version'])
def version_cli_flag(request):
return request.param
def test_cli_group_version_option(version_cli_flag):
result = runner.invoke(main, [version_cli_flag])
assert result.exit_code == 0
assert result.output == 'cibopath, version {}\n'.format(__version__)
|
import pytest
from cibopath import __version__
@pytest.fixture(params=['-V', '--version'])
def version_cli_flag(request):
return request.param
def test_cli_group_version_option(cli_runner, version_cli_flag):
result = cli_runner([version_cli_flag])
assert result.exit_code == 0
assert result.output == 'cibopath, version {}\n'.format(__version__)
| Use cli_runner fixture in test | Use cli_runner fixture in test
| Python | bsd-3-clause | hackebrot/cibopath |
- from click.testing import CliRunner
import pytest
- from cibopath.cli import main
from cibopath import __version__
-
- runner = CliRunner()
@pytest.fixture(params=['-V', '--version'])
def version_cli_flag(request):
return request.param
- def test_cli_group_version_option(version_cli_flag):
+ def test_cli_group_version_option(cli_runner, version_cli_flag):
- result = runner.invoke(main, [version_cli_flag])
+ result = cli_runner([version_cli_flag])
assert result.exit_code == 0
assert result.output == 'cibopath, version {}\n'.format(__version__)
| Use cli_runner fixture in test | ## Code Before:
from click.testing import CliRunner
import pytest
from cibopath.cli import main
from cibopath import __version__
runner = CliRunner()
@pytest.fixture(params=['-V', '--version'])
def version_cli_flag(request):
return request.param
def test_cli_group_version_option(version_cli_flag):
result = runner.invoke(main, [version_cli_flag])
assert result.exit_code == 0
assert result.output == 'cibopath, version {}\n'.format(__version__)
## Instruction:
Use cli_runner fixture in test
## Code After:
import pytest
from cibopath import __version__
@pytest.fixture(params=['-V', '--version'])
def version_cli_flag(request):
return request.param
def test_cli_group_version_option(cli_runner, version_cli_flag):
result = cli_runner([version_cli_flag])
assert result.exit_code == 0
assert result.output == 'cibopath, version {}\n'.format(__version__)
|
f814e945d3e62c87c5f86ef5ac37c5feb733b83d | tests/test_ext.py | tests/test_ext.py | from __future__ import absolute_import, unicode_literals
import unittest
from mopidy import config, ext
class ExtensionTest(unittest.TestCase):
def setUp(self): # noqa: N802
self.ext = ext.Extension()
def test_dist_name_is_none(self):
self.assertIsNone(self.ext.dist_name)
def test_ext_name_is_none(self):
self.assertIsNone(self.ext.ext_name)
def test_version_is_none(self):
self.assertIsNone(self.ext.version)
def test_get_default_config_raises_not_implemented(self):
with self.assertRaises(NotImplementedError):
self.ext.get_default_config()
def test_get_config_schema_returns_extension_schema(self):
schema = self.ext.get_config_schema()
self.assertIsInstance(schema['enabled'], config.Boolean)
def test_validate_environment_does_nothing_by_default(self):
self.assertIsNone(self.ext.validate_environment())
def test_setup_raises_not_implemented(self):
with self.assertRaises(NotImplementedError):
self.ext.setup(None)
| from __future__ import absolute_import, unicode_literals
import pytest
from mopidy import config, ext
@pytest.fixture
def extension():
return ext.Extension()
def test_dist_name_is_none(extension):
assert extension.dist_name is None
def test_ext_name_is_none(extension):
assert extension.ext_name is None
def test_version_is_none(extension):
assert extension.version is None
def test_get_default_config_raises_not_implemented(extension):
with pytest.raises(NotImplementedError):
extension.get_default_config()
def test_get_config_schema_returns_extension_schema(extension):
schema = extension.get_config_schema()
assert isinstance(schema['enabled'], config.Boolean)
def test_validate_environment_does_nothing_by_default(extension):
assert extension.validate_environment() is None
def test_setup_raises_not_implemented(extension):
with pytest.raises(NotImplementedError):
extension.setup(None)
| Convert ext test to pytests | tests: Convert ext test to pytests
| Python | apache-2.0 | mokieyue/mopidy,bencevans/mopidy,ZenithDK/mopidy,jodal/mopidy,quartz55/mopidy,pacificIT/mopidy,pacificIT/mopidy,quartz55/mopidy,ali/mopidy,swak/mopidy,tkem/mopidy,bencevans/mopidy,mopidy/mopidy,SuperStarPL/mopidy,dbrgn/mopidy,hkariti/mopidy,glogiotatidis/mopidy,mokieyue/mopidy,ali/mopidy,bacontext/mopidy,glogiotatidis/mopidy,diandiankan/mopidy,dbrgn/mopidy,mopidy/mopidy,jmarsik/mopidy,glogiotatidis/mopidy,bacontext/mopidy,vrs01/mopidy,kingosticks/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,vrs01/mopidy,dbrgn/mopidy,jcass77/mopidy,glogiotatidis/mopidy,swak/mopidy,bencevans/mopidy,mokieyue/mopidy,adamcik/mopidy,tkem/mopidy,rawdlite/mopidy,bacontext/mopidy,pacificIT/mopidy,ali/mopidy,jcass77/mopidy,rawdlite/mopidy,mopidy/mopidy,vrs01/mopidy,tkem/mopidy,dbrgn/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,tkem/mopidy,jodal/mopidy,diandiankan/mopidy,jmarsik/mopidy,quartz55/mopidy,ZenithDK/mopidy,jodal/mopidy,ZenithDK/mopidy,jmarsik/mopidy,swak/mopidy,ZenithDK/mopidy,bacontext/mopidy,kingosticks/mopidy,quartz55/mopidy,pacificIT/mopidy,SuperStarPL/mopidy,hkariti/mopidy,ali/mopidy,hkariti/mopidy,adamcik/mopidy,vrs01/mopidy,rawdlite/mopidy,hkariti/mopidy,rawdlite/mopidy,mokieyue/mopidy,adamcik/mopidy,swak/mopidy,kingosticks/mopidy,bencevans/mopidy,jmarsik/mopidy,jcass77/mopidy | from __future__ import absolute_import, unicode_literals
- import unittest
+ import pytest
from mopidy import config, ext
- class ExtensionTest(unittest.TestCase):
+ @pytest.fixture
+ def extension():
+ return ext.Extension()
- def setUp(self): # noqa: N802
- self.ext = ext.Extension()
- def test_dist_name_is_none(self):
+ def test_dist_name_is_none(extension):
- self.assertIsNone(self.ext.dist_name)
+ assert extension.dist_name is None
- def test_ext_name_is_none(self):
- self.assertIsNone(self.ext.ext_name)
- def test_version_is_none(self):
- self.assertIsNone(self.ext.version)
+ def test_ext_name_is_none(extension):
+ assert extension.ext_name is None
- def test_get_default_config_raises_not_implemented(self):
- with self.assertRaises(NotImplementedError):
- self.ext.get_default_config()
+ def test_version_is_none(extension):
+ assert extension.version is None
- def test_get_config_schema_returns_extension_schema(self):
- schema = self.ext.get_config_schema()
- self.assertIsInstance(schema['enabled'], config.Boolean)
- def test_validate_environment_does_nothing_by_default(self):
- self.assertIsNone(self.ext.validate_environment())
- def test_setup_raises_not_implemented(self):
+ def test_get_default_config_raises_not_implemented(extension):
- with self.assertRaises(NotImplementedError):
+ with pytest.raises(NotImplementedError):
- self.ext.setup(None)
+ extension.get_default_config()
+
+ def test_get_config_schema_returns_extension_schema(extension):
+ schema = extension.get_config_schema()
+ assert isinstance(schema['enabled'], config.Boolean)
+
+
+ def test_validate_environment_does_nothing_by_default(extension):
+ assert extension.validate_environment() is None
+
+
+ def test_setup_raises_not_implemented(extension):
+ with pytest.raises(NotImplementedError):
+ extension.setup(None)
+ | Convert ext test to pytests | ## Code Before:
from __future__ import absolute_import, unicode_literals
import unittest
from mopidy import config, ext
class ExtensionTest(unittest.TestCase):
def setUp(self): # noqa: N802
self.ext = ext.Extension()
def test_dist_name_is_none(self):
self.assertIsNone(self.ext.dist_name)
def test_ext_name_is_none(self):
self.assertIsNone(self.ext.ext_name)
def test_version_is_none(self):
self.assertIsNone(self.ext.version)
def test_get_default_config_raises_not_implemented(self):
with self.assertRaises(NotImplementedError):
self.ext.get_default_config()
def test_get_config_schema_returns_extension_schema(self):
schema = self.ext.get_config_schema()
self.assertIsInstance(schema['enabled'], config.Boolean)
def test_validate_environment_does_nothing_by_default(self):
self.assertIsNone(self.ext.validate_environment())
def test_setup_raises_not_implemented(self):
with self.assertRaises(NotImplementedError):
self.ext.setup(None)
## Instruction:
Convert ext test to pytests
## Code After:
from __future__ import absolute_import, unicode_literals
import pytest
from mopidy import config, ext
@pytest.fixture
def extension():
return ext.Extension()
def test_dist_name_is_none(extension):
assert extension.dist_name is None
def test_ext_name_is_none(extension):
assert extension.ext_name is None
def test_version_is_none(extension):
assert extension.version is None
def test_get_default_config_raises_not_implemented(extension):
with pytest.raises(NotImplementedError):
extension.get_default_config()
def test_get_config_schema_returns_extension_schema(extension):
schema = extension.get_config_schema()
assert isinstance(schema['enabled'], config.Boolean)
def test_validate_environment_does_nothing_by_default(extension):
assert extension.validate_environment() is None
def test_setup_raises_not_implemented(extension):
with pytest.raises(NotImplementedError):
extension.setup(None)
|
27ffcae96c5dce976517035b25a5c72f10e2ec99 | tool_spatialdb.py | tool_spatialdb.py |
import os
import sys
import eol_scons
tools = ['sqlitedb','doxygen','prefixoptions']
env = Environment(tools = ['default'] + tools)
platform = env['PLATFORM']
thisdir = env.Dir('.').srcnode().abspath
# define the tool
def spatialdb(env):
env.AppendUnique(CPPPATH =[thisdir,])
env.AppendLibrary('spatialdb')
env.AppendLibrary('geos')
env.AppendLibrary('geos_c')
env.AppendLibrary('proj')
if (platform != 'posix'):
env.AppendLibrary('iconv')
env.Replace(CCFLAGS=['-g','-O2'])
env.Require(tools)
Export('spatialdb')
# build the SpatialDB library
libsources = Split("""
SpatiaLiteDB.cpp
""")
headers = Split("""
SpatiaLiteDB.h
""")
libspatialdb = env.Library('spatialdb', libsources)
env.Default(libspatialdb)
html = env.Apidocs(libsources + headers, DOXYFILE_DICT={'PROJECT_NAME':'SpatialDB', 'PROJECT_NUMBER':'1.0'})
|
import os
import sys
import eol_scons
tools = ['sqlitedb','doxygen','prefixoptions']
env = Environment(tools = ['default'] + tools)
platform = env['PLATFORM']
thisdir = env.Dir('.').srcnode().abspath
# define the tool
def spatialdb(env):
env.AppendUnique(CPPPATH =[thisdir,])
env.AppendLibrary('spatialdb')
env.AppendLibrary('geos')
env.AppendLibrary('geos_c')
env.AppendLibrary('proj')
if (platform != 'posix'):
env.AppendLibrary('iconv')
env.Require(tools)
Export('spatialdb')
# build the SpatialDB library
libsources = Split("""
SpatiaLiteDB.cpp
""")
headers = Split("""
SpatiaLiteDB.h
""")
libspatialdb = env.Library('spatialdb', libsources)
env.Default(libspatialdb)
html = env.Apidocs(libsources + headers, DOXYFILE_DICT={'PROJECT_NAME':'SpatialDB', 'PROJECT_NUMBER':'1.0'})
| Use GLOBAL_TOOLs rather than Export/Import for project wide configuration. | Use GLOBAL_TOOLs rather than Export/Import for project wide configuration.
| Python | bsd-3-clause | ncareol/spatialdb,ncareol/spatialdb |
import os
import sys
import eol_scons
tools = ['sqlitedb','doxygen','prefixoptions']
env = Environment(tools = ['default'] + tools)
platform = env['PLATFORM']
thisdir = env.Dir('.').srcnode().abspath
# define the tool
def spatialdb(env):
env.AppendUnique(CPPPATH =[thisdir,])
env.AppendLibrary('spatialdb')
env.AppendLibrary('geos')
env.AppendLibrary('geos_c')
env.AppendLibrary('proj')
if (platform != 'posix'):
env.AppendLibrary('iconv')
- env.Replace(CCFLAGS=['-g','-O2'])
env.Require(tools)
Export('spatialdb')
# build the SpatialDB library
libsources = Split("""
SpatiaLiteDB.cpp
""")
headers = Split("""
SpatiaLiteDB.h
""")
libspatialdb = env.Library('spatialdb', libsources)
env.Default(libspatialdb)
html = env.Apidocs(libsources + headers, DOXYFILE_DICT={'PROJECT_NAME':'SpatialDB', 'PROJECT_NUMBER':'1.0'})
| Use GLOBAL_TOOLs rather than Export/Import for project wide configuration. | ## Code Before:
import os
import sys
import eol_scons
tools = ['sqlitedb','doxygen','prefixoptions']
env = Environment(tools = ['default'] + tools)
platform = env['PLATFORM']
thisdir = env.Dir('.').srcnode().abspath
# define the tool
def spatialdb(env):
env.AppendUnique(CPPPATH =[thisdir,])
env.AppendLibrary('spatialdb')
env.AppendLibrary('geos')
env.AppendLibrary('geos_c')
env.AppendLibrary('proj')
if (platform != 'posix'):
env.AppendLibrary('iconv')
env.Replace(CCFLAGS=['-g','-O2'])
env.Require(tools)
Export('spatialdb')
# build the SpatialDB library
libsources = Split("""
SpatiaLiteDB.cpp
""")
headers = Split("""
SpatiaLiteDB.h
""")
libspatialdb = env.Library('spatialdb', libsources)
env.Default(libspatialdb)
html = env.Apidocs(libsources + headers, DOXYFILE_DICT={'PROJECT_NAME':'SpatialDB', 'PROJECT_NUMBER':'1.0'})
## Instruction:
Use GLOBAL_TOOLs rather than Export/Import for project wide configuration.
## Code After:
import os
import sys
import eol_scons
tools = ['sqlitedb','doxygen','prefixoptions']
env = Environment(tools = ['default'] + tools)
platform = env['PLATFORM']
thisdir = env.Dir('.').srcnode().abspath
# define the tool
def spatialdb(env):
env.AppendUnique(CPPPATH =[thisdir,])
env.AppendLibrary('spatialdb')
env.AppendLibrary('geos')
env.AppendLibrary('geos_c')
env.AppendLibrary('proj')
if (platform != 'posix'):
env.AppendLibrary('iconv')
env.Require(tools)
Export('spatialdb')
# build the SpatialDB library
libsources = Split("""
SpatiaLiteDB.cpp
""")
headers = Split("""
SpatiaLiteDB.h
""")
libspatialdb = env.Library('spatialdb', libsources)
env.Default(libspatialdb)
html = env.Apidocs(libsources + headers, DOXYFILE_DICT={'PROJECT_NAME':'SpatialDB', 'PROJECT_NUMBER':'1.0'})
|
da59d7481668a7133eebcd12b4d5ecfb655296a6 | test/test_blob_filter.py | test/test_blob_filter.py | """Test the blob filter."""
from pathlib import Path
from typing import Sequence, Tuple
from unittest.mock import MagicMock
import pytest
from git.index.typ import BlobFilter, StageType
from git.objects import Blob
from git.types import PathLike
# fmt: off
@pytest.mark.parametrize('paths, stage_type, path, expected_result', [
((Path("foo"),), 0, Path("foo"), True),
((Path("foo"),), 0, Path("foo/bar"), True),
((Path("foo/bar"),), 0, Path("foo"), False),
((Path("foo"), Path("bar")), 0, Path("foo"), True),
])
# fmt: on
def test_blob_filter(paths: Sequence[PathLike], stage_type: StageType, path: PathLike, expected_result: bool) -> None:
"""Test the blob filter."""
blob_filter = BlobFilter(paths)
binsha = MagicMock(__len__=lambda self: 20)
blob: Blob = Blob(repo=MagicMock(), binsha=binsha, path=path)
stage_blob: Tuple[StageType, Blob] = (stage_type, blob)
result = blob_filter(stage_blob)
assert result == expected_result
| """Test the blob filter."""
from pathlib import Path
from typing import Sequence, Tuple
from unittest.mock import MagicMock
import pytest
from git.index.typ import BlobFilter, StageType
from git.objects import Blob
from git.types import PathLike
# fmt: off
@pytest.mark.parametrize('paths, path, expected_result', [
((Path("foo"),), Path("foo"), True),
((Path("foo"),), Path("foo/bar"), True),
((Path("foo/bar"),), Path("foo"), False),
((Path("foo"), Path("bar")), Path("foo"), True),
])
# fmt: on
def test_blob_filter(paths: Sequence[PathLike], path: PathLike, expected_result: bool) -> None:
"""Test the blob filter."""
blob_filter = BlobFilter(paths)
binsha = MagicMock(__len__=lambda self: 20)
stage_type: StageType = 0
blob: Blob = Blob(repo=MagicMock(), binsha=binsha, path=path)
stage_blob: Tuple[StageType, Blob] = (stage_type, blob)
result = blob_filter(stage_blob)
assert result == expected_result
| Remove stage type as parameter from blob filter test | Remove stage type as parameter from blob filter test
| Python | bsd-3-clause | gitpython-developers/GitPython,gitpython-developers/gitpython,gitpython-developers/GitPython,gitpython-developers/gitpython | """Test the blob filter."""
from pathlib import Path
from typing import Sequence, Tuple
from unittest.mock import MagicMock
import pytest
from git.index.typ import BlobFilter, StageType
from git.objects import Blob
from git.types import PathLike
# fmt: off
- @pytest.mark.parametrize('paths, stage_type, path, expected_result', [
+ @pytest.mark.parametrize('paths, path, expected_result', [
- ((Path("foo"),), 0, Path("foo"), True),
+ ((Path("foo"),), Path("foo"), True),
- ((Path("foo"),), 0, Path("foo/bar"), True),
+ ((Path("foo"),), Path("foo/bar"), True),
- ((Path("foo/bar"),), 0, Path("foo"), False),
+ ((Path("foo/bar"),), Path("foo"), False),
- ((Path("foo"), Path("bar")), 0, Path("foo"), True),
+ ((Path("foo"), Path("bar")), Path("foo"), True),
])
# fmt: on
- def test_blob_filter(paths: Sequence[PathLike], stage_type: StageType, path: PathLike, expected_result: bool) -> None:
+ def test_blob_filter(paths: Sequence[PathLike], path: PathLike, expected_result: bool) -> None:
"""Test the blob filter."""
blob_filter = BlobFilter(paths)
binsha = MagicMock(__len__=lambda self: 20)
+ stage_type: StageType = 0
blob: Blob = Blob(repo=MagicMock(), binsha=binsha, path=path)
stage_blob: Tuple[StageType, Blob] = (stage_type, blob)
result = blob_filter(stage_blob)
assert result == expected_result
| Remove stage type as parameter from blob filter test | ## Code Before:
"""Test the blob filter."""
from pathlib import Path
from typing import Sequence, Tuple
from unittest.mock import MagicMock
import pytest
from git.index.typ import BlobFilter, StageType
from git.objects import Blob
from git.types import PathLike
# fmt: off
@pytest.mark.parametrize('paths, stage_type, path, expected_result', [
((Path("foo"),), 0, Path("foo"), True),
((Path("foo"),), 0, Path("foo/bar"), True),
((Path("foo/bar"),), 0, Path("foo"), False),
((Path("foo"), Path("bar")), 0, Path("foo"), True),
])
# fmt: on
def test_blob_filter(paths: Sequence[PathLike], stage_type: StageType, path: PathLike, expected_result: bool) -> None:
"""Test the blob filter."""
blob_filter = BlobFilter(paths)
binsha = MagicMock(__len__=lambda self: 20)
blob: Blob = Blob(repo=MagicMock(), binsha=binsha, path=path)
stage_blob: Tuple[StageType, Blob] = (stage_type, blob)
result = blob_filter(stage_blob)
assert result == expected_result
## Instruction:
Remove stage type as parameter from blob filter test
## Code After:
"""Test the blob filter."""
from pathlib import Path
from typing import Sequence, Tuple
from unittest.mock import MagicMock
import pytest
from git.index.typ import BlobFilter, StageType
from git.objects import Blob
from git.types import PathLike
# fmt: off
@pytest.mark.parametrize('paths, path, expected_result', [
((Path("foo"),), Path("foo"), True),
((Path("foo"),), Path("foo/bar"), True),
((Path("foo/bar"),), Path("foo"), False),
((Path("foo"), Path("bar")), Path("foo"), True),
])
# fmt: on
def test_blob_filter(paths: Sequence[PathLike], path: PathLike, expected_result: bool) -> None:
"""Test the blob filter."""
blob_filter = BlobFilter(paths)
binsha = MagicMock(__len__=lambda self: 20)
stage_type: StageType = 0
blob: Blob = Blob(repo=MagicMock(), binsha=binsha, path=path)
stage_blob: Tuple[StageType, Blob] = (stage_type, blob)
result = blob_filter(stage_blob)
assert result == expected_result
|
431ca4f2d44656ef9f97be50718712c6f3a0fa9b | qtawesome/tests/test_qtawesome.py | qtawesome/tests/test_qtawesome.py | # Standard library imports
import subprocess
# Test Library imports
import pytest
# Local imports
import qtawesome as qta
from qtawesome.iconic_font import IconicFont
def test_segfault_import():
output_number = subprocess.call('python -c "import qtawesome '
'; qtawesome.icon()"', shell=True)
assert output_number == 0
def test_unique_font_family_name(qtbot):
"""
Test that each font used by qtawesome has a unique name. If this test
fails, this probably means that you need to rename the family name of
some fonts. Please see PR #98 for more details on why it is necessary and
on how to do this.
Regression test for Issue #107
"""
resource = qta._instance()
assert isinstance(resource, IconicFont)
prefixes = list(resource.fontname.keys())
assert prefixes
fontnames = set(resource.fontname.values())
assert fontnames
assert len(prefixes) == len(fontnames)
if __name__ == "__main__":
pytest.main()
| # Standard library imports
import subprocess
import collections
# Test Library imports
import pytest
# Local imports
import qtawesome as qta
from qtawesome.iconic_font import IconicFont
def test_segfault_import():
output_number = subprocess.call('python -c "import qtawesome '
'; qtawesome.icon()"', shell=True)
assert output_number == 0
def test_unique_font_family_name(qtbot):
"""
Test that each font used by qtawesome has a unique name. If this test
fails, this probably means that you need to rename the family name of
some fonts. Please see PR #98 for more details on why it is necessary and
on how to do this.
Regression test for Issue #107
"""
resource = qta._instance()
assert isinstance(resource, IconicFont)
# Check that the fonts were loaded successfully.
fontnames = resource.fontname.values()
assert fontnames
# Check that qtawesome does not load fonts with duplicate family names.
duplicates = [fontname for fontname, count in
collections.Counter(fontnames).items() if count > 1]
assert not duplicates
if __name__ == "__main__":
pytest.main()
| Make the test more comprehensive. | Make the test more comprehensive.
| Python | mit | spyder-ide/qtawesome | # Standard library imports
import subprocess
+ import collections
# Test Library imports
import pytest
# Local imports
import qtawesome as qta
from qtawesome.iconic_font import IconicFont
def test_segfault_import():
output_number = subprocess.call('python -c "import qtawesome '
'; qtawesome.icon()"', shell=True)
assert output_number == 0
def test_unique_font_family_name(qtbot):
"""
Test that each font used by qtawesome has a unique name. If this test
fails, this probably means that you need to rename the family name of
some fonts. Please see PR #98 for more details on why it is necessary and
on how to do this.
Regression test for Issue #107
"""
resource = qta._instance()
assert isinstance(resource, IconicFont)
+ # Check that the fonts were loaded successfully.
- prefixes = list(resource.fontname.keys())
- assert prefixes
-
- fontnames = set(resource.fontname.values())
+ fontnames = resource.fontname.values()
assert fontnames
- assert len(prefixes) == len(fontnames)
+ # Check that qtawesome does not load fonts with duplicate family names.
+ duplicates = [fontname for fontname, count in
+ collections.Counter(fontnames).items() if count > 1]
+ assert not duplicates
if __name__ == "__main__":
pytest.main()
| Make the test more comprehensive. | ## Code Before:
# Standard library imports
import subprocess
# Test Library imports
import pytest
# Local imports
import qtawesome as qta
from qtawesome.iconic_font import IconicFont
def test_segfault_import():
output_number = subprocess.call('python -c "import qtawesome '
'; qtawesome.icon()"', shell=True)
assert output_number == 0
def test_unique_font_family_name(qtbot):
"""
Test that each font used by qtawesome has a unique name. If this test
fails, this probably means that you need to rename the family name of
some fonts. Please see PR #98 for more details on why it is necessary and
on how to do this.
Regression test for Issue #107
"""
resource = qta._instance()
assert isinstance(resource, IconicFont)
prefixes = list(resource.fontname.keys())
assert prefixes
fontnames = set(resource.fontname.values())
assert fontnames
assert len(prefixes) == len(fontnames)
if __name__ == "__main__":
pytest.main()
## Instruction:
Make the test more comprehensive.
## Code After:
# Standard library imports
import subprocess
import collections
# Test Library imports
import pytest
# Local imports
import qtawesome as qta
from qtawesome.iconic_font import IconicFont
def test_segfault_import():
output_number = subprocess.call('python -c "import qtawesome '
'; qtawesome.icon()"', shell=True)
assert output_number == 0
def test_unique_font_family_name(qtbot):
"""
Test that each font used by qtawesome has a unique name. If this test
fails, this probably means that you need to rename the family name of
some fonts. Please see PR #98 for more details on why it is necessary and
on how to do this.
Regression test for Issue #107
"""
resource = qta._instance()
assert isinstance(resource, IconicFont)
# Check that the fonts were loaded successfully.
fontnames = resource.fontname.values()
assert fontnames
# Check that qtawesome does not load fonts with duplicate family names.
duplicates = [fontname for fontname, count in
collections.Counter(fontnames).items() if count > 1]
assert not duplicates
if __name__ == "__main__":
pytest.main()
|
3d8ec94e61735b84c3b24b44d79fcd57611a93ee | mndeps.py | mndeps.py | from mininet.topo import ( SingleSwitchTopo, LinearTopo,
SingleSwitchReversedTopo )
from mininet.topolib import TreeTopo
from mininet.util import buildTopo
import psycopg2
TOPOS = { 'linear': LinearTopo,
'reversed': SingleSwitchReversedTopo,
'single': SingleSwitchTopo,
'tree': TreeTopo,
'torus': TorusTopo }
def build(opts):
return buildTopo(TOPOS, opts)
| from mininet.topo import ( SingleSwitchTopo, LinearTopo,
SingleSwitchReversedTopo )
from mininet.topolib import TreeTopo
from mininet.util import buildTopo
import psycopg2
TOPOS = { 'linear': LinearTopo,
'reversed': SingleSwitchReversedTopo,
'single': SingleSwitchTopo,
'tree': TreeTopo
}
def build(opts):
return buildTopo(TOPOS, opts)
| Fix bug after removing torus | Fix bug after removing torus
| Python | apache-2.0 | ravel-net/ravel,ravel-net/ravel | from mininet.topo import ( SingleSwitchTopo, LinearTopo,
SingleSwitchReversedTopo )
from mininet.topolib import TreeTopo
from mininet.util import buildTopo
import psycopg2
TOPOS = { 'linear': LinearTopo,
'reversed': SingleSwitchReversedTopo,
'single': SingleSwitchTopo,
- 'tree': TreeTopo,
+ 'tree': TreeTopo
- 'torus': TorusTopo }
+ }
def build(opts):
return buildTopo(TOPOS, opts)
| Fix bug after removing torus | ## Code Before:
from mininet.topo import ( SingleSwitchTopo, LinearTopo,
SingleSwitchReversedTopo )
from mininet.topolib import TreeTopo
from mininet.util import buildTopo
import psycopg2
TOPOS = { 'linear': LinearTopo,
'reversed': SingleSwitchReversedTopo,
'single': SingleSwitchTopo,
'tree': TreeTopo,
'torus': TorusTopo }
def build(opts):
return buildTopo(TOPOS, opts)
## Instruction:
Fix bug after removing torus
## Code After:
from mininet.topo import ( SingleSwitchTopo, LinearTopo,
SingleSwitchReversedTopo )
from mininet.topolib import TreeTopo
from mininet.util import buildTopo
import psycopg2
TOPOS = { 'linear': LinearTopo,
'reversed': SingleSwitchReversedTopo,
'single': SingleSwitchTopo,
'tree': TreeTopo
}
def build(opts):
return buildTopo(TOPOS, opts)
|