repo
stringlengths 7
55
| path
stringlengths 4
127
| func_name
stringlengths 1
88
| original_string
stringlengths 75
19.8k
| language
stringclasses 1
value | code
stringlengths 75
19.8k
| code_tokens
sequence | docstring
stringlengths 3
17.3k
| docstring_tokens
sequence | sha
stringlengths 40
40
| url
stringlengths 87
242
| partition
stringclasses 1
value |
---|---|---|---|---|---|---|---|---|---|---|---|
tchellomello/python-amcrest | src/amcrest/http.py | Http._set_name | def _set_name(self):
"""Set device name."""
try:
self._name = pretty(self.machine_name)
self._serial = self.serial_number
except AttributeError:
self._name = None
self._serial = None | python | def _set_name(self):
"""Set device name."""
try:
self._name = pretty(self.machine_name)
self._serial = self.serial_number
except AttributeError:
self._name = None
self._serial = None | [
"def",
"_set_name",
"(",
"self",
")",
":",
"try",
":",
"self",
".",
"_name",
"=",
"pretty",
"(",
"self",
".",
"machine_name",
")",
"self",
".",
"_serial",
"=",
"self",
".",
"serial_number",
"except",
"AttributeError",
":",
"self",
".",
"_name",
"=",
"None",
"self",
".",
"_serial",
"=",
"None"
] | Set device name. | [
"Set",
"device",
"name",
"."
] | ed842139e234de2eaf6ee8fb480214711cde1249 | https://github.com/tchellomello/python-amcrest/blob/ed842139e234de2eaf6ee8fb480214711cde1249/src/amcrest/http.py#L101-L108 | train |
tchellomello/python-amcrest | src/amcrest/utils.py | to_unit | def to_unit(value, unit='B'):
"""Convert bytes to give unit."""
byte_array = ['B', 'KB', 'MB', 'GB', 'TB']
if not isinstance(value, (int, float)):
value = float(value)
if unit in byte_array:
result = value / 1024**byte_array.index(unit)
return round(result, PRECISION), unit
return value | python | def to_unit(value, unit='B'):
"""Convert bytes to give unit."""
byte_array = ['B', 'KB', 'MB', 'GB', 'TB']
if not isinstance(value, (int, float)):
value = float(value)
if unit in byte_array:
result = value / 1024**byte_array.index(unit)
return round(result, PRECISION), unit
return value | [
"def",
"to_unit",
"(",
"value",
",",
"unit",
"=",
"'B'",
")",
":",
"byte_array",
"=",
"[",
"'B'",
",",
"'KB'",
",",
"'MB'",
",",
"'GB'",
",",
"'TB'",
"]",
"if",
"not",
"isinstance",
"(",
"value",
",",
"(",
"int",
",",
"float",
")",
")",
":",
"value",
"=",
"float",
"(",
"value",
")",
"if",
"unit",
"in",
"byte_array",
":",
"result",
"=",
"value",
"/",
"1024",
"**",
"byte_array",
".",
"index",
"(",
"unit",
")",
"return",
"round",
"(",
"result",
",",
"PRECISION",
")",
",",
"unit",
"return",
"value"
] | Convert bytes to give unit. | [
"Convert",
"bytes",
"to",
"give",
"unit",
"."
] | ed842139e234de2eaf6ee8fb480214711cde1249 | https://github.com/tchellomello/python-amcrest/blob/ed842139e234de2eaf6ee8fb480214711cde1249/src/amcrest/utils.py#L57-L68 | train |
tchellomello/python-amcrest | src/amcrest/special.py | Special.realtime_stream | def realtime_stream(self, channel=1, typeno=0, path_file=None):
"""
If the stream is redirect to a file, use mplayer tool to
visualize the video record
camera.realtime_stream(path_file="/home/user/Desktop/myvideo)
$ mplayer /home/user/Desktop/myvideo
"""
ret = self.command(
'realmonitor.cgi?action=getStream&channel={0}&subtype={1}'.format(
channel, typeno)
)
if path_file:
with open(path_file, 'wb') as out_file:
shutil.copyfileobj(ret.raw, out_file)
return ret.raw | python | def realtime_stream(self, channel=1, typeno=0, path_file=None):
"""
If the stream is redirect to a file, use mplayer tool to
visualize the video record
camera.realtime_stream(path_file="/home/user/Desktop/myvideo)
$ mplayer /home/user/Desktop/myvideo
"""
ret = self.command(
'realmonitor.cgi?action=getStream&channel={0}&subtype={1}'.format(
channel, typeno)
)
if path_file:
with open(path_file, 'wb') as out_file:
shutil.copyfileobj(ret.raw, out_file)
return ret.raw | [
"def",
"realtime_stream",
"(",
"self",
",",
"channel",
"=",
"1",
",",
"typeno",
"=",
"0",
",",
"path_file",
"=",
"None",
")",
":",
"ret",
"=",
"self",
".",
"command",
"(",
"'realmonitor.cgi?action=getStream&channel={0}&subtype={1}'",
".",
"format",
"(",
"channel",
",",
"typeno",
")",
")",
"if",
"path_file",
":",
"with",
"open",
"(",
"path_file",
",",
"'wb'",
")",
"as",
"out_file",
":",
"shutil",
".",
"copyfileobj",
"(",
"ret",
".",
"raw",
",",
"out_file",
")",
"return",
"ret",
".",
"raw"
] | If the stream is redirect to a file, use mplayer tool to
visualize the video record
camera.realtime_stream(path_file="/home/user/Desktop/myvideo)
$ mplayer /home/user/Desktop/myvideo | [
"If",
"the",
"stream",
"is",
"redirect",
"to",
"a",
"file",
"use",
"mplayer",
"tool",
"to",
"visualize",
"the",
"video",
"record"
] | ed842139e234de2eaf6ee8fb480214711cde1249 | https://github.com/tchellomello/python-amcrest/blob/ed842139e234de2eaf6ee8fb480214711cde1249/src/amcrest/special.py#L20-L37 | train |
tchellomello/python-amcrest | src/amcrest/special.py | Special.rtsp_url | def rtsp_url(self, channelno=None, typeno=None):
"""
Return RTSP streaming url
Params:
channelno: integer, the video channel index which starts from 1,
default 1 if not specified.
typeno: the stream type, default 0 if not specified. It can be
the following value:
0-Main Stream
1-Extra Stream 1 (Sub Stream)
2-Extra Stream 2 (Sub Stream)
"""
if channelno is None:
channelno = 1
if typeno is None:
typeno = 0
cmd = 'cam/realmonitor?channel={0}&subtype={1}'.format(
channelno, typeno)
try:
port = ':' + [x.split('=')[1] for x in self.rtsp_config.split()
if x.startswith('table.RTSP.Port=')][0]
except IndexError:
port = ''
return 'rtsp://{}:{}@{}{}/{}'.format(
self._user, self._password, self._host, port, cmd) | python | def rtsp_url(self, channelno=None, typeno=None):
"""
Return RTSP streaming url
Params:
channelno: integer, the video channel index which starts from 1,
default 1 if not specified.
typeno: the stream type, default 0 if not specified. It can be
the following value:
0-Main Stream
1-Extra Stream 1 (Sub Stream)
2-Extra Stream 2 (Sub Stream)
"""
if channelno is None:
channelno = 1
if typeno is None:
typeno = 0
cmd = 'cam/realmonitor?channel={0}&subtype={1}'.format(
channelno, typeno)
try:
port = ':' + [x.split('=')[1] for x in self.rtsp_config.split()
if x.startswith('table.RTSP.Port=')][0]
except IndexError:
port = ''
return 'rtsp://{}:{}@{}{}/{}'.format(
self._user, self._password, self._host, port, cmd) | [
"def",
"rtsp_url",
"(",
"self",
",",
"channelno",
"=",
"None",
",",
"typeno",
"=",
"None",
")",
":",
"if",
"channelno",
"is",
"None",
":",
"channelno",
"=",
"1",
"if",
"typeno",
"is",
"None",
":",
"typeno",
"=",
"0",
"cmd",
"=",
"'cam/realmonitor?channel={0}&subtype={1}'",
".",
"format",
"(",
"channelno",
",",
"typeno",
")",
"try",
":",
"port",
"=",
"':'",
"+",
"[",
"x",
".",
"split",
"(",
"'='",
")",
"[",
"1",
"]",
"for",
"x",
"in",
"self",
".",
"rtsp_config",
".",
"split",
"(",
")",
"if",
"x",
".",
"startswith",
"(",
"'table.RTSP.Port='",
")",
"]",
"[",
"0",
"]",
"except",
"IndexError",
":",
"port",
"=",
"''",
"return",
"'rtsp://{}:{}@{}{}/{}'",
".",
"format",
"(",
"self",
".",
"_user",
",",
"self",
".",
"_password",
",",
"self",
".",
"_host",
",",
"port",
",",
"cmd",
")"
] | Return RTSP streaming url
Params:
channelno: integer, the video channel index which starts from 1,
default 1 if not specified.
typeno: the stream type, default 0 if not specified. It can be
the following value:
0-Main Stream
1-Extra Stream 1 (Sub Stream)
2-Extra Stream 2 (Sub Stream) | [
"Return",
"RTSP",
"streaming",
"url"
] | ed842139e234de2eaf6ee8fb480214711cde1249 | https://github.com/tchellomello/python-amcrest/blob/ed842139e234de2eaf6ee8fb480214711cde1249/src/amcrest/special.py#L39-L69 | train |
tchellomello/python-amcrest | src/amcrest/special.py | Special.mjpeg_url | def mjpeg_url(self, channelno=None, typeno=None):
"""
Return MJPEG streaming url
Params:
channelno: integer, the video channel index which starts from 1,
default 1 if not specified.
typeno: the stream type, default 0 if not specified. It can be
the following value:
0-Main Stream
1-Extra Stream 1 (Sub Stream)
2-Extra Stream 2 (Sub Stream)
"""
if channelno is None:
channelno = 0
if typeno is None:
typeno = 1
cmd = "mjpg/video.cgi?channel={0}&subtype={1}".format(
channelno, typeno)
return '{0}{1}'.format(self._base_url, cmd) | python | def mjpeg_url(self, channelno=None, typeno=None):
"""
Return MJPEG streaming url
Params:
channelno: integer, the video channel index which starts from 1,
default 1 if not specified.
typeno: the stream type, default 0 if not specified. It can be
the following value:
0-Main Stream
1-Extra Stream 1 (Sub Stream)
2-Extra Stream 2 (Sub Stream)
"""
if channelno is None:
channelno = 0
if typeno is None:
typeno = 1
cmd = "mjpg/video.cgi?channel={0}&subtype={1}".format(
channelno, typeno)
return '{0}{1}'.format(self._base_url, cmd) | [
"def",
"mjpeg_url",
"(",
"self",
",",
"channelno",
"=",
"None",
",",
"typeno",
"=",
"None",
")",
":",
"if",
"channelno",
"is",
"None",
":",
"channelno",
"=",
"0",
"if",
"typeno",
"is",
"None",
":",
"typeno",
"=",
"1",
"cmd",
"=",
"\"mjpg/video.cgi?channel={0}&subtype={1}\"",
".",
"format",
"(",
"channelno",
",",
"typeno",
")",
"return",
"'{0}{1}'",
".",
"format",
"(",
"self",
".",
"_base_url",
",",
"cmd",
")"
] | Return MJPEG streaming url
Params:
channelno: integer, the video channel index which starts from 1,
default 1 if not specified.
typeno: the stream type, default 0 if not specified. It can be
the following value:
0-Main Stream
1-Extra Stream 1 (Sub Stream)
2-Extra Stream 2 (Sub Stream) | [
"Return",
"MJPEG",
"streaming",
"url"
] | ed842139e234de2eaf6ee8fb480214711cde1249 | https://github.com/tchellomello/python-amcrest/blob/ed842139e234de2eaf6ee8fb480214711cde1249/src/amcrest/special.py#L95-L118 | train |
tchellomello/python-amcrest | src/amcrest/network.py | Network.scan_devices | def scan_devices(self, subnet, timeout=None):
"""
Scan cameras in a range of ips
Params:
subnet - subnet, i.e: 192.168.1.0/24
if mask not used, assuming mask 24
timeout_sec - timeout in sec
Returns:
"""
# Maximum range from mask
# Format is mask: max_range
max_range = {
16: 256,
24: 256,
25: 128,
27: 32,
28: 16,
29: 8,
30: 4,
31: 2
}
# If user didn't provide mask, use /24
if "/" not in subnet:
mask = int(24)
network = subnet
else:
network, mask = subnet.split("/")
mask = int(mask)
if mask not in max_range:
raise RuntimeError("Cannot determine the subnet mask!")
# Default logic is remove everything from last "." to the end
# This logic change in case mask is 16
network = network.rpartition(".")[0]
if mask == 16:
# For mask 16, we must cut the last two
# entries with .
# pylint: disable=unused-variable
for i in range(0, 1):
network = network.rpartition(".")[0]
# Trigger the scan
# For clear coding, let's keep the logic in if/else (mask16)
# instead of only one if
if mask == 16:
for seq1 in range(0, max_range[mask]):
for seq2 in range(0, max_range[mask]):
ipaddr = "{0}.{1}.{2}".format(network, seq1, seq2)
thd = threading.Thread(
target=self.__raw_scan, args=(ipaddr, timeout)
)
thd.start()
else:
for seq1 in range(0, max_range[mask]):
ipaddr = "{0}.{1}".format(network, seq1)
thd = threading.Thread(
target=self.__raw_scan, args=(ipaddr, timeout)
)
thd.start()
return self.amcrest_ips | python | def scan_devices(self, subnet, timeout=None):
"""
Scan cameras in a range of ips
Params:
subnet - subnet, i.e: 192.168.1.0/24
if mask not used, assuming mask 24
timeout_sec - timeout in sec
Returns:
"""
# Maximum range from mask
# Format is mask: max_range
max_range = {
16: 256,
24: 256,
25: 128,
27: 32,
28: 16,
29: 8,
30: 4,
31: 2
}
# If user didn't provide mask, use /24
if "/" not in subnet:
mask = int(24)
network = subnet
else:
network, mask = subnet.split("/")
mask = int(mask)
if mask not in max_range:
raise RuntimeError("Cannot determine the subnet mask!")
# Default logic is remove everything from last "." to the end
# This logic change in case mask is 16
network = network.rpartition(".")[0]
if mask == 16:
# For mask 16, we must cut the last two
# entries with .
# pylint: disable=unused-variable
for i in range(0, 1):
network = network.rpartition(".")[0]
# Trigger the scan
# For clear coding, let's keep the logic in if/else (mask16)
# instead of only one if
if mask == 16:
for seq1 in range(0, max_range[mask]):
for seq2 in range(0, max_range[mask]):
ipaddr = "{0}.{1}.{2}".format(network, seq1, seq2)
thd = threading.Thread(
target=self.__raw_scan, args=(ipaddr, timeout)
)
thd.start()
else:
for seq1 in range(0, max_range[mask]):
ipaddr = "{0}.{1}".format(network, seq1)
thd = threading.Thread(
target=self.__raw_scan, args=(ipaddr, timeout)
)
thd.start()
return self.amcrest_ips | [
"def",
"scan_devices",
"(",
"self",
",",
"subnet",
",",
"timeout",
"=",
"None",
")",
":",
"# Maximum range from mask",
"# Format is mask: max_range",
"max_range",
"=",
"{",
"16",
":",
"256",
",",
"24",
":",
"256",
",",
"25",
":",
"128",
",",
"27",
":",
"32",
",",
"28",
":",
"16",
",",
"29",
":",
"8",
",",
"30",
":",
"4",
",",
"31",
":",
"2",
"}",
"# If user didn't provide mask, use /24",
"if",
"\"/\"",
"not",
"in",
"subnet",
":",
"mask",
"=",
"int",
"(",
"24",
")",
"network",
"=",
"subnet",
"else",
":",
"network",
",",
"mask",
"=",
"subnet",
".",
"split",
"(",
"\"/\"",
")",
"mask",
"=",
"int",
"(",
"mask",
")",
"if",
"mask",
"not",
"in",
"max_range",
":",
"raise",
"RuntimeError",
"(",
"\"Cannot determine the subnet mask!\"",
")",
"# Default logic is remove everything from last \".\" to the end",
"# This logic change in case mask is 16",
"network",
"=",
"network",
".",
"rpartition",
"(",
"\".\"",
")",
"[",
"0",
"]",
"if",
"mask",
"==",
"16",
":",
"# For mask 16, we must cut the last two",
"# entries with .",
"# pylint: disable=unused-variable",
"for",
"i",
"in",
"range",
"(",
"0",
",",
"1",
")",
":",
"network",
"=",
"network",
".",
"rpartition",
"(",
"\".\"",
")",
"[",
"0",
"]",
"# Trigger the scan",
"# For clear coding, let's keep the logic in if/else (mask16)",
"# instead of only one if",
"if",
"mask",
"==",
"16",
":",
"for",
"seq1",
"in",
"range",
"(",
"0",
",",
"max_range",
"[",
"mask",
"]",
")",
":",
"for",
"seq2",
"in",
"range",
"(",
"0",
",",
"max_range",
"[",
"mask",
"]",
")",
":",
"ipaddr",
"=",
"\"{0}.{1}.{2}\"",
".",
"format",
"(",
"network",
",",
"seq1",
",",
"seq2",
")",
"thd",
"=",
"threading",
".",
"Thread",
"(",
"target",
"=",
"self",
".",
"__raw_scan",
",",
"args",
"=",
"(",
"ipaddr",
",",
"timeout",
")",
")",
"thd",
".",
"start",
"(",
")",
"else",
":",
"for",
"seq1",
"in",
"range",
"(",
"0",
",",
"max_range",
"[",
"mask",
"]",
")",
":",
"ipaddr",
"=",
"\"{0}.{1}\"",
".",
"format",
"(",
"network",
",",
"seq1",
")",
"thd",
"=",
"threading",
".",
"Thread",
"(",
"target",
"=",
"self",
".",
"__raw_scan",
",",
"args",
"=",
"(",
"ipaddr",
",",
"timeout",
")",
")",
"thd",
".",
"start",
"(",
")",
"return",
"self",
".",
"amcrest_ips"
] | Scan cameras in a range of ips
Params:
subnet - subnet, i.e: 192.168.1.0/24
if mask not used, assuming mask 24
timeout_sec - timeout in sec
Returns: | [
"Scan",
"cameras",
"in",
"a",
"range",
"of",
"ips"
] | ed842139e234de2eaf6ee8fb480214711cde1249 | https://github.com/tchellomello/python-amcrest/blob/ed842139e234de2eaf6ee8fb480214711cde1249/src/amcrest/network.py#L41-L109 | train |
peerplays-network/python-peerplays | peerplays/peerplays.py | PeerPlays.disallow | def disallow(
self, foreign, permission="active", account=None, threshold=None, **kwargs
):
""" Remove additional access to an account by some other public
key or account.
:param str foreign: The foreign account that will obtain access
:param str permission: (optional) The actual permission to
modify (defaults to ``active``)
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
:param int threshold: The threshold that needs to be reached
by signatures to be able to interact
"""
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
if permission not in ["owner", "active"]:
raise ValueError("Permission needs to be either 'owner', or 'active")
account = Account(account, blockchain_instance=self)
authority = account[permission]
try:
pubkey = PublicKey(foreign, prefix=self.prefix)
affected_items = list(
filter(lambda x: x[0] == str(pubkey), authority["key_auths"])
)
authority["key_auths"] = list(
filter(lambda x: x[0] != str(pubkey), authority["key_auths"])
)
except:
try:
foreign_account = Account(foreign, blockchain_instance=self)
affected_items = list(
filter(
lambda x: x[0] == foreign_account["id"],
authority["account_auths"],
)
)
authority["account_auths"] = list(
filter(
lambda x: x[0] != foreign_account["id"],
authority["account_auths"],
)
)
except:
raise ValueError("Unknown foreign account or unvalid public key")
if not affected_items:
raise ValueError("Changes nothing!")
removed_weight = affected_items[0][1]
# Define threshold
if threshold:
authority["weight_threshold"] = threshold
# Correct threshold (at most by the amount removed from the
# authority)
try:
self._test_weights_treshold(authority)
except:
log.critical(
"The account's threshold will be reduced by %d" % (removed_weight)
)
authority["weight_threshold"] -= removed_weight
self._test_weights_treshold(authority)
op = operations.Account_update(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"account": account["id"],
permission: authority,
"extensions": {},
}
)
if permission == "owner":
return self.finalizeOp(op, account["name"], "owner", **kwargs)
else:
return self.finalizeOp(op, account["name"], "active", **kwargs) | python | def disallow(
self, foreign, permission="active", account=None, threshold=None, **kwargs
):
""" Remove additional access to an account by some other public
key or account.
:param str foreign: The foreign account that will obtain access
:param str permission: (optional) The actual permission to
modify (defaults to ``active``)
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
:param int threshold: The threshold that needs to be reached
by signatures to be able to interact
"""
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
if permission not in ["owner", "active"]:
raise ValueError("Permission needs to be either 'owner', or 'active")
account = Account(account, blockchain_instance=self)
authority = account[permission]
try:
pubkey = PublicKey(foreign, prefix=self.prefix)
affected_items = list(
filter(lambda x: x[0] == str(pubkey), authority["key_auths"])
)
authority["key_auths"] = list(
filter(lambda x: x[0] != str(pubkey), authority["key_auths"])
)
except:
try:
foreign_account = Account(foreign, blockchain_instance=self)
affected_items = list(
filter(
lambda x: x[0] == foreign_account["id"],
authority["account_auths"],
)
)
authority["account_auths"] = list(
filter(
lambda x: x[0] != foreign_account["id"],
authority["account_auths"],
)
)
except:
raise ValueError("Unknown foreign account or unvalid public key")
if not affected_items:
raise ValueError("Changes nothing!")
removed_weight = affected_items[0][1]
# Define threshold
if threshold:
authority["weight_threshold"] = threshold
# Correct threshold (at most by the amount removed from the
# authority)
try:
self._test_weights_treshold(authority)
except:
log.critical(
"The account's threshold will be reduced by %d" % (removed_weight)
)
authority["weight_threshold"] -= removed_weight
self._test_weights_treshold(authority)
op = operations.Account_update(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"account": account["id"],
permission: authority,
"extensions": {},
}
)
if permission == "owner":
return self.finalizeOp(op, account["name"], "owner", **kwargs)
else:
return self.finalizeOp(op, account["name"], "active", **kwargs) | [
"def",
"disallow",
"(",
"self",
",",
"foreign",
",",
"permission",
"=",
"\"active\"",
",",
"account",
"=",
"None",
",",
"threshold",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"not",
"account",
":",
"if",
"\"default_account\"",
"in",
"self",
".",
"config",
":",
"account",
"=",
"self",
".",
"config",
"[",
"\"default_account\"",
"]",
"if",
"not",
"account",
":",
"raise",
"ValueError",
"(",
"\"You need to provide an account\"",
")",
"if",
"permission",
"not",
"in",
"[",
"\"owner\"",
",",
"\"active\"",
"]",
":",
"raise",
"ValueError",
"(",
"\"Permission needs to be either 'owner', or 'active\"",
")",
"account",
"=",
"Account",
"(",
"account",
",",
"blockchain_instance",
"=",
"self",
")",
"authority",
"=",
"account",
"[",
"permission",
"]",
"try",
":",
"pubkey",
"=",
"PublicKey",
"(",
"foreign",
",",
"prefix",
"=",
"self",
".",
"prefix",
")",
"affected_items",
"=",
"list",
"(",
"filter",
"(",
"lambda",
"x",
":",
"x",
"[",
"0",
"]",
"==",
"str",
"(",
"pubkey",
")",
",",
"authority",
"[",
"\"key_auths\"",
"]",
")",
")",
"authority",
"[",
"\"key_auths\"",
"]",
"=",
"list",
"(",
"filter",
"(",
"lambda",
"x",
":",
"x",
"[",
"0",
"]",
"!=",
"str",
"(",
"pubkey",
")",
",",
"authority",
"[",
"\"key_auths\"",
"]",
")",
")",
"except",
":",
"try",
":",
"foreign_account",
"=",
"Account",
"(",
"foreign",
",",
"blockchain_instance",
"=",
"self",
")",
"affected_items",
"=",
"list",
"(",
"filter",
"(",
"lambda",
"x",
":",
"x",
"[",
"0",
"]",
"==",
"foreign_account",
"[",
"\"id\"",
"]",
",",
"authority",
"[",
"\"account_auths\"",
"]",
",",
")",
")",
"authority",
"[",
"\"account_auths\"",
"]",
"=",
"list",
"(",
"filter",
"(",
"lambda",
"x",
":",
"x",
"[",
"0",
"]",
"!=",
"foreign_account",
"[",
"\"id\"",
"]",
",",
"authority",
"[",
"\"account_auths\"",
"]",
",",
")",
")",
"except",
":",
"raise",
"ValueError",
"(",
"\"Unknown foreign account or unvalid public key\"",
")",
"if",
"not",
"affected_items",
":",
"raise",
"ValueError",
"(",
"\"Changes nothing!\"",
")",
"removed_weight",
"=",
"affected_items",
"[",
"0",
"]",
"[",
"1",
"]",
"# Define threshold",
"if",
"threshold",
":",
"authority",
"[",
"\"weight_threshold\"",
"]",
"=",
"threshold",
"# Correct threshold (at most by the amount removed from the",
"# authority)",
"try",
":",
"self",
".",
"_test_weights_treshold",
"(",
"authority",
")",
"except",
":",
"log",
".",
"critical",
"(",
"\"The account's threshold will be reduced by %d\"",
"%",
"(",
"removed_weight",
")",
")",
"authority",
"[",
"\"weight_threshold\"",
"]",
"-=",
"removed_weight",
"self",
".",
"_test_weights_treshold",
"(",
"authority",
")",
"op",
"=",
"operations",
".",
"Account_update",
"(",
"*",
"*",
"{",
"\"fee\"",
":",
"{",
"\"amount\"",
":",
"0",
",",
"\"asset_id\"",
":",
"\"1.3.0\"",
"}",
",",
"\"account\"",
":",
"account",
"[",
"\"id\"",
"]",
",",
"permission",
":",
"authority",
",",
"\"extensions\"",
":",
"{",
"}",
",",
"}",
")",
"if",
"permission",
"==",
"\"owner\"",
":",
"return",
"self",
".",
"finalizeOp",
"(",
"op",
",",
"account",
"[",
"\"name\"",
"]",
",",
"\"owner\"",
",",
"*",
"*",
"kwargs",
")",
"else",
":",
"return",
"self",
".",
"finalizeOp",
"(",
"op",
",",
"account",
"[",
"\"name\"",
"]",
",",
"\"active\"",
",",
"*",
"*",
"kwargs",
")"
] | Remove additional access to an account by some other public
key or account.
:param str foreign: The foreign account that will obtain access
:param str permission: (optional) The actual permission to
modify (defaults to ``active``)
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
:param int threshold: The threshold that needs to be reached
by signatures to be able to interact | [
"Remove",
"additional",
"access",
"to",
"an",
"account",
"by",
"some",
"other",
"public",
"key",
"or",
"account",
"."
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplays/peerplays.py#L439-L520 | train |
peerplays-network/python-peerplays | peerplays/peerplays.py | PeerPlays.approvewitness | def approvewitness(self, witnesses, account=None, **kwargs):
""" Approve a witness
:param list witnesses: list of Witness name or id
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
"""
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account, blockchain_instance=self)
options = account["options"]
if not isinstance(witnesses, (list, set, tuple)):
witnesses = {witnesses}
for witness in witnesses:
witness = Witness(witness, blockchain_instance=self)
options["votes"].append(witness["vote_id"])
options["votes"] = list(set(options["votes"]))
options["num_witness"] = len(
list(filter(lambda x: float(x.split(":")[0]) == 1, options["votes"]))
)
op = operations.Account_update(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"account": account["id"],
"new_options": options,
"extensions": {},
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | python | def approvewitness(self, witnesses, account=None, **kwargs):
""" Approve a witness
:param list witnesses: list of Witness name or id
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
"""
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account, blockchain_instance=self)
options = account["options"]
if not isinstance(witnesses, (list, set, tuple)):
witnesses = {witnesses}
for witness in witnesses:
witness = Witness(witness, blockchain_instance=self)
options["votes"].append(witness["vote_id"])
options["votes"] = list(set(options["votes"]))
options["num_witness"] = len(
list(filter(lambda x: float(x.split(":")[0]) == 1, options["votes"]))
)
op = operations.Account_update(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"account": account["id"],
"new_options": options,
"extensions": {},
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | [
"def",
"approvewitness",
"(",
"self",
",",
"witnesses",
",",
"account",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"not",
"account",
":",
"if",
"\"default_account\"",
"in",
"self",
".",
"config",
":",
"account",
"=",
"self",
".",
"config",
"[",
"\"default_account\"",
"]",
"if",
"not",
"account",
":",
"raise",
"ValueError",
"(",
"\"You need to provide an account\"",
")",
"account",
"=",
"Account",
"(",
"account",
",",
"blockchain_instance",
"=",
"self",
")",
"options",
"=",
"account",
"[",
"\"options\"",
"]",
"if",
"not",
"isinstance",
"(",
"witnesses",
",",
"(",
"list",
",",
"set",
",",
"tuple",
")",
")",
":",
"witnesses",
"=",
"{",
"witnesses",
"}",
"for",
"witness",
"in",
"witnesses",
":",
"witness",
"=",
"Witness",
"(",
"witness",
",",
"blockchain_instance",
"=",
"self",
")",
"options",
"[",
"\"votes\"",
"]",
".",
"append",
"(",
"witness",
"[",
"\"vote_id\"",
"]",
")",
"options",
"[",
"\"votes\"",
"]",
"=",
"list",
"(",
"set",
"(",
"options",
"[",
"\"votes\"",
"]",
")",
")",
"options",
"[",
"\"num_witness\"",
"]",
"=",
"len",
"(",
"list",
"(",
"filter",
"(",
"lambda",
"x",
":",
"float",
"(",
"x",
".",
"split",
"(",
"\":\"",
")",
"[",
"0",
"]",
")",
"==",
"1",
",",
"options",
"[",
"\"votes\"",
"]",
")",
")",
")",
"op",
"=",
"operations",
".",
"Account_update",
"(",
"*",
"*",
"{",
"\"fee\"",
":",
"{",
"\"amount\"",
":",
"0",
",",
"\"asset_id\"",
":",
"\"1.3.0\"",
"}",
",",
"\"account\"",
":",
"account",
"[",
"\"id\"",
"]",
",",
"\"new_options\"",
":",
"options",
",",
"\"extensions\"",
":",
"{",
"}",
",",
"\"prefix\"",
":",
"self",
".",
"prefix",
",",
"}",
")",
"return",
"self",
".",
"finalizeOp",
"(",
"op",
",",
"account",
"[",
"\"name\"",
"]",
",",
"\"active\"",
",",
"*",
"*",
"kwargs",
")"
] | Approve a witness
:param list witnesses: list of Witness name or id
:param str account: (optional) the account to allow access
to (defaults to ``default_account``) | [
"Approve",
"a",
"witness"
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplays/peerplays.py#L556-L592 | train |
peerplays-network/python-peerplays | peerplays/peerplays.py | PeerPlays.approvecommittee | def approvecommittee(self, committees, account=None, **kwargs):
""" Approve a committee
:param list committees: list of committee member name or id
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
"""
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account, blockchain_instance=self)
options = account["options"]
if not isinstance(committees, (list, set, tuple)):
committees = {committees}
for committee in committees:
committee = Committee(committee, blockchain_instance=self)
options["votes"].append(committee["vote_id"])
options["votes"] = list(set(options["votes"]))
options["num_committee"] = len(
list(filter(lambda x: float(x.split(":")[0]) == 0, options["votes"]))
)
op = operations.Account_update(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"account": account["id"],
"new_options": options,
"extensions": {},
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | python | def approvecommittee(self, committees, account=None, **kwargs):
""" Approve a committee
:param list committees: list of committee member name or id
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
"""
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account, blockchain_instance=self)
options = account["options"]
if not isinstance(committees, (list, set, tuple)):
committees = {committees}
for committee in committees:
committee = Committee(committee, blockchain_instance=self)
options["votes"].append(committee["vote_id"])
options["votes"] = list(set(options["votes"]))
options["num_committee"] = len(
list(filter(lambda x: float(x.split(":")[0]) == 0, options["votes"]))
)
op = operations.Account_update(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"account": account["id"],
"new_options": options,
"extensions": {},
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | [
"def",
"approvecommittee",
"(",
"self",
",",
"committees",
",",
"account",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"not",
"account",
":",
"if",
"\"default_account\"",
"in",
"self",
".",
"config",
":",
"account",
"=",
"self",
".",
"config",
"[",
"\"default_account\"",
"]",
"if",
"not",
"account",
":",
"raise",
"ValueError",
"(",
"\"You need to provide an account\"",
")",
"account",
"=",
"Account",
"(",
"account",
",",
"blockchain_instance",
"=",
"self",
")",
"options",
"=",
"account",
"[",
"\"options\"",
"]",
"if",
"not",
"isinstance",
"(",
"committees",
",",
"(",
"list",
",",
"set",
",",
"tuple",
")",
")",
":",
"committees",
"=",
"{",
"committees",
"}",
"for",
"committee",
"in",
"committees",
":",
"committee",
"=",
"Committee",
"(",
"committee",
",",
"blockchain_instance",
"=",
"self",
")",
"options",
"[",
"\"votes\"",
"]",
".",
"append",
"(",
"committee",
"[",
"\"vote_id\"",
"]",
")",
"options",
"[",
"\"votes\"",
"]",
"=",
"list",
"(",
"set",
"(",
"options",
"[",
"\"votes\"",
"]",
")",
")",
"options",
"[",
"\"num_committee\"",
"]",
"=",
"len",
"(",
"list",
"(",
"filter",
"(",
"lambda",
"x",
":",
"float",
"(",
"x",
".",
"split",
"(",
"\":\"",
")",
"[",
"0",
"]",
")",
"==",
"0",
",",
"options",
"[",
"\"votes\"",
"]",
")",
")",
")",
"op",
"=",
"operations",
".",
"Account_update",
"(",
"*",
"*",
"{",
"\"fee\"",
":",
"{",
"\"amount\"",
":",
"0",
",",
"\"asset_id\"",
":",
"\"1.3.0\"",
"}",
",",
"\"account\"",
":",
"account",
"[",
"\"id\"",
"]",
",",
"\"new_options\"",
":",
"options",
",",
"\"extensions\"",
":",
"{",
"}",
",",
"\"prefix\"",
":",
"self",
".",
"prefix",
",",
"}",
")",
"return",
"self",
".",
"finalizeOp",
"(",
"op",
",",
"account",
"[",
"\"name\"",
"]",
",",
"\"active\"",
",",
"*",
"*",
"kwargs",
")"
] | Approve a committee
:param list committees: list of committee member name or id
:param str account: (optional) the account to allow access
to (defaults to ``default_account``) | [
"Approve",
"a",
"committee"
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplays/peerplays.py#L633-L669 | train |
peerplays-network/python-peerplays | peerplays/peerplays.py | PeerPlays.betting_market_rules_create | def betting_market_rules_create(self, names, descriptions, account=None, **kwargs):
""" Create betting market rules
:param list names: Internationalized names, e.g. ``[['de', 'Foo'],
['en', 'bar']]``
:param list descriptions: Internationalized descriptions, e.g.
``[['de', 'Foo'], ['en', 'bar']]``
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
"""
assert isinstance(names, list)
assert isinstance(descriptions, list)
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account)
op = operations.Betting_market_rules_create(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"name": names,
"description": descriptions,
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | python | def betting_market_rules_create(self, names, descriptions, account=None, **kwargs):
""" Create betting market rules
:param list names: Internationalized names, e.g. ``[['de', 'Foo'],
['en', 'bar']]``
:param list descriptions: Internationalized descriptions, e.g.
``[['de', 'Foo'], ['en', 'bar']]``
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
"""
assert isinstance(names, list)
assert isinstance(descriptions, list)
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account)
op = operations.Betting_market_rules_create(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"name": names,
"description": descriptions,
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | [
"def",
"betting_market_rules_create",
"(",
"self",
",",
"names",
",",
"descriptions",
",",
"account",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"assert",
"isinstance",
"(",
"names",
",",
"list",
")",
"assert",
"isinstance",
"(",
"descriptions",
",",
"list",
")",
"if",
"not",
"account",
":",
"if",
"\"default_account\"",
"in",
"self",
".",
"config",
":",
"account",
"=",
"self",
".",
"config",
"[",
"\"default_account\"",
"]",
"if",
"not",
"account",
":",
"raise",
"ValueError",
"(",
"\"You need to provide an account\"",
")",
"account",
"=",
"Account",
"(",
"account",
")",
"op",
"=",
"operations",
".",
"Betting_market_rules_create",
"(",
"*",
"*",
"{",
"\"fee\"",
":",
"{",
"\"amount\"",
":",
"0",
",",
"\"asset_id\"",
":",
"\"1.3.0\"",
"}",
",",
"\"name\"",
":",
"names",
",",
"\"description\"",
":",
"descriptions",
",",
"\"prefix\"",
":",
"self",
".",
"prefix",
",",
"}",
")",
"return",
"self",
".",
"finalizeOp",
"(",
"op",
",",
"account",
"[",
"\"name\"",
"]",
",",
"\"active\"",
",",
"*",
"*",
"kwargs",
")"
] | Create betting market rules
:param list names: Internationalized names, e.g. ``[['de', 'Foo'],
['en', 'bar']]``
:param list descriptions: Internationalized descriptions, e.g.
``[['de', 'Foo'], ['en', 'bar']]``
:param str account: (optional) the account to allow access
to (defaults to ``default_account``) | [
"Create",
"betting",
"market",
"rules"
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplays/peerplays.py#L1149-L1176 | train |
peerplays-network/python-peerplays | peerplays/peerplays.py | PeerPlays.betting_market_rules_update | def betting_market_rules_update(
self, rules_id, names, descriptions, account=None, **kwargs
):
""" Update betting market rules
:param str rules_id: Id of the betting market rules to update
:param list names: Internationalized names, e.g. ``[['de', 'Foo'],
['en', 'bar']]``
:param list descriptions: Internationalized descriptions, e.g.
``[['de', 'Foo'], ['en', 'bar']]``
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
"""
assert isinstance(names, list)
assert isinstance(descriptions, list)
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account)
rule = Rule(rules_id)
op = operations.Betting_market_rules_update(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"betting_market_rules_id": rule["id"],
"new_name": names,
"new_description": descriptions,
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | python | def betting_market_rules_update(
self, rules_id, names, descriptions, account=None, **kwargs
):
""" Update betting market rules
:param str rules_id: Id of the betting market rules to update
:param list names: Internationalized names, e.g. ``[['de', 'Foo'],
['en', 'bar']]``
:param list descriptions: Internationalized descriptions, e.g.
``[['de', 'Foo'], ['en', 'bar']]``
:param str account: (optional) the account to allow access
to (defaults to ``default_account``)
"""
assert isinstance(names, list)
assert isinstance(descriptions, list)
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account)
rule = Rule(rules_id)
op = operations.Betting_market_rules_update(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"betting_market_rules_id": rule["id"],
"new_name": names,
"new_description": descriptions,
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | [
"def",
"betting_market_rules_update",
"(",
"self",
",",
"rules_id",
",",
"names",
",",
"descriptions",
",",
"account",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"assert",
"isinstance",
"(",
"names",
",",
"list",
")",
"assert",
"isinstance",
"(",
"descriptions",
",",
"list",
")",
"if",
"not",
"account",
":",
"if",
"\"default_account\"",
"in",
"self",
".",
"config",
":",
"account",
"=",
"self",
".",
"config",
"[",
"\"default_account\"",
"]",
"if",
"not",
"account",
":",
"raise",
"ValueError",
"(",
"\"You need to provide an account\"",
")",
"account",
"=",
"Account",
"(",
"account",
")",
"rule",
"=",
"Rule",
"(",
"rules_id",
")",
"op",
"=",
"operations",
".",
"Betting_market_rules_update",
"(",
"*",
"*",
"{",
"\"fee\"",
":",
"{",
"\"amount\"",
":",
"0",
",",
"\"asset_id\"",
":",
"\"1.3.0\"",
"}",
",",
"\"betting_market_rules_id\"",
":",
"rule",
"[",
"\"id\"",
"]",
",",
"\"new_name\"",
":",
"names",
",",
"\"new_description\"",
":",
"descriptions",
",",
"\"prefix\"",
":",
"self",
".",
"prefix",
",",
"}",
")",
"return",
"self",
".",
"finalizeOp",
"(",
"op",
",",
"account",
"[",
"\"name\"",
"]",
",",
"\"active\"",
",",
"*",
"*",
"kwargs",
")"
] | Update betting market rules
:param str rules_id: Id of the betting market rules to update
:param list names: Internationalized names, e.g. ``[['de', 'Foo'],
['en', 'bar']]``
:param list descriptions: Internationalized descriptions, e.g.
``[['de', 'Foo'], ['en', 'bar']]``
:param str account: (optional) the account to allow access
to (defaults to ``default_account``) | [
"Update",
"betting",
"market",
"rules"
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplays/peerplays.py#L1178-L1210 | train |
peerplays-network/python-peerplays | peerplays/peerplays.py | PeerPlays.bet_place | def bet_place(
self,
betting_market_id,
amount_to_bet,
backer_multiplier,
back_or_lay,
account=None,
**kwargs
):
""" Place a bet
:param str betting_market_id: The identifier for the market to bet
in
:param peerplays.amount.Amount amount_to_bet: Amount to bet with
:param int backer_multiplier: Multipler for backer
:param str back_or_lay: "back" or "lay" the bet
:param str account: (optional) the account to bet (defaults
to ``default_account``)
"""
from . import GRAPHENE_BETTING_ODDS_PRECISION
assert isinstance(amount_to_bet, Amount)
assert back_or_lay in ["back", "lay"]
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account)
bm = BettingMarket(betting_market_id)
op = operations.Bet_place(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"bettor_id": account["id"],
"betting_market_id": bm["id"],
"amount_to_bet": amount_to_bet.json(),
"backer_multiplier": (
int(backer_multiplier * GRAPHENE_BETTING_ODDS_PRECISION)
),
"back_or_lay": back_or_lay,
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | python | def bet_place(
self,
betting_market_id,
amount_to_bet,
backer_multiplier,
back_or_lay,
account=None,
**kwargs
):
""" Place a bet
:param str betting_market_id: The identifier for the market to bet
in
:param peerplays.amount.Amount amount_to_bet: Amount to bet with
:param int backer_multiplier: Multipler for backer
:param str back_or_lay: "back" or "lay" the bet
:param str account: (optional) the account to bet (defaults
to ``default_account``)
"""
from . import GRAPHENE_BETTING_ODDS_PRECISION
assert isinstance(amount_to_bet, Amount)
assert back_or_lay in ["back", "lay"]
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account)
bm = BettingMarket(betting_market_id)
op = operations.Bet_place(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"bettor_id": account["id"],
"betting_market_id": bm["id"],
"amount_to_bet": amount_to_bet.json(),
"backer_multiplier": (
int(backer_multiplier * GRAPHENE_BETTING_ODDS_PRECISION)
),
"back_or_lay": back_or_lay,
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | [
"def",
"bet_place",
"(",
"self",
",",
"betting_market_id",
",",
"amount_to_bet",
",",
"backer_multiplier",
",",
"back_or_lay",
",",
"account",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"from",
".",
"import",
"GRAPHENE_BETTING_ODDS_PRECISION",
"assert",
"isinstance",
"(",
"amount_to_bet",
",",
"Amount",
")",
"assert",
"back_or_lay",
"in",
"[",
"\"back\"",
",",
"\"lay\"",
"]",
"if",
"not",
"account",
":",
"if",
"\"default_account\"",
"in",
"self",
".",
"config",
":",
"account",
"=",
"self",
".",
"config",
"[",
"\"default_account\"",
"]",
"if",
"not",
"account",
":",
"raise",
"ValueError",
"(",
"\"You need to provide an account\"",
")",
"account",
"=",
"Account",
"(",
"account",
")",
"bm",
"=",
"BettingMarket",
"(",
"betting_market_id",
")",
"op",
"=",
"operations",
".",
"Bet_place",
"(",
"*",
"*",
"{",
"\"fee\"",
":",
"{",
"\"amount\"",
":",
"0",
",",
"\"asset_id\"",
":",
"\"1.3.0\"",
"}",
",",
"\"bettor_id\"",
":",
"account",
"[",
"\"id\"",
"]",
",",
"\"betting_market_id\"",
":",
"bm",
"[",
"\"id\"",
"]",
",",
"\"amount_to_bet\"",
":",
"amount_to_bet",
".",
"json",
"(",
")",
",",
"\"backer_multiplier\"",
":",
"(",
"int",
"(",
"backer_multiplier",
"*",
"GRAPHENE_BETTING_ODDS_PRECISION",
")",
")",
",",
"\"back_or_lay\"",
":",
"back_or_lay",
",",
"\"prefix\"",
":",
"self",
".",
"prefix",
",",
"}",
")",
"return",
"self",
".",
"finalizeOp",
"(",
"op",
",",
"account",
"[",
"\"name\"",
"]",
",",
"\"active\"",
",",
"*",
"*",
"kwargs",
")"
] | Place a bet
:param str betting_market_id: The identifier for the market to bet
in
:param peerplays.amount.Amount amount_to_bet: Amount to bet with
:param int backer_multiplier: Multipler for backer
:param str back_or_lay: "back" or "lay" the bet
:param str account: (optional) the account to bet (defaults
to ``default_account``) | [
"Place",
"a",
"bet"
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplays/peerplays.py#L1488-L1531 | train |
peerplays-network/python-peerplays | peerplays/peerplays.py | PeerPlays.bet_cancel | def bet_cancel(self, bet_to_cancel, account=None, **kwargs):
""" Cancel a bet
:param str bet_to_cancel: The identifier that identifies the bet to
cancel
:param str account: (optional) the account that owns the bet
(defaults to ``default_account``)
"""
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account)
bet = Bet(bet_to_cancel)
op = operations.Bet_cancel(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"bettor_id": account["id"],
"bet_to_cancel": bet["id"],
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | python | def bet_cancel(self, bet_to_cancel, account=None, **kwargs):
""" Cancel a bet
:param str bet_to_cancel: The identifier that identifies the bet to
cancel
:param str account: (optional) the account that owns the bet
(defaults to ``default_account``)
"""
if not account:
if "default_account" in self.config:
account = self.config["default_account"]
if not account:
raise ValueError("You need to provide an account")
account = Account(account)
bet = Bet(bet_to_cancel)
op = operations.Bet_cancel(
**{
"fee": {"amount": 0, "asset_id": "1.3.0"},
"bettor_id": account["id"],
"bet_to_cancel": bet["id"],
"prefix": self.prefix,
}
)
return self.finalizeOp(op, account["name"], "active", **kwargs) | [
"def",
"bet_cancel",
"(",
"self",
",",
"bet_to_cancel",
",",
"account",
"=",
"None",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"not",
"account",
":",
"if",
"\"default_account\"",
"in",
"self",
".",
"config",
":",
"account",
"=",
"self",
".",
"config",
"[",
"\"default_account\"",
"]",
"if",
"not",
"account",
":",
"raise",
"ValueError",
"(",
"\"You need to provide an account\"",
")",
"account",
"=",
"Account",
"(",
"account",
")",
"bet",
"=",
"Bet",
"(",
"bet_to_cancel",
")",
"op",
"=",
"operations",
".",
"Bet_cancel",
"(",
"*",
"*",
"{",
"\"fee\"",
":",
"{",
"\"amount\"",
":",
"0",
",",
"\"asset_id\"",
":",
"\"1.3.0\"",
"}",
",",
"\"bettor_id\"",
":",
"account",
"[",
"\"id\"",
"]",
",",
"\"bet_to_cancel\"",
":",
"bet",
"[",
"\"id\"",
"]",
",",
"\"prefix\"",
":",
"self",
".",
"prefix",
",",
"}",
")",
"return",
"self",
".",
"finalizeOp",
"(",
"op",
",",
"account",
"[",
"\"name\"",
"]",
",",
"\"active\"",
",",
"*",
"*",
"kwargs",
")"
] | Cancel a bet
:param str bet_to_cancel: The identifier that identifies the bet to
cancel
:param str account: (optional) the account that owns the bet
(defaults to ``default_account``) | [
"Cancel",
"a",
"bet"
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplays/peerplays.py#L1533-L1556 | train |
peerplays-network/python-peerplays | peerplays/cli/decorators.py | verbose | def verbose(f):
""" Add verbose flags and add logging handlers
"""
@click.pass_context
def new_func(ctx, *args, **kwargs):
global log
verbosity = ["critical", "error", "warn", "info", "debug"][
int(min(ctx.obj.get("verbose", 0), 4))
]
log.setLevel(getattr(logging, verbosity.upper()))
formatter = logging.Formatter(
"%(asctime)s - %(name)s - %(levelname)s - %(message)s"
)
ch = logging.StreamHandler()
ch.setLevel(getattr(logging, verbosity.upper()))
ch.setFormatter(formatter)
log.addHandler(ch)
# GrapheneAPI logging
if ctx.obj.get("verbose", 0) > 4:
verbosity = ["critical", "error", "warn", "info", "debug"][
int(min(ctx.obj.get("verbose", 4) - 4, 4))
]
log = logging.getLogger("grapheneapi")
log.setLevel(getattr(logging, verbosity.upper()))
log.addHandler(ch)
if ctx.obj.get("verbose", 0) > 8:
verbosity = ["critical", "error", "warn", "info", "debug"][
int(min(ctx.obj.get("verbose", 8) - 8, 4))
]
log = logging.getLogger("graphenebase")
log.setLevel(getattr(logging, verbosity.upper()))
log.addHandler(ch)
return ctx.invoke(f, *args, **kwargs)
return update_wrapper(new_func, f) | python | def verbose(f):
""" Add verbose flags and add logging handlers
"""
@click.pass_context
def new_func(ctx, *args, **kwargs):
global log
verbosity = ["critical", "error", "warn", "info", "debug"][
int(min(ctx.obj.get("verbose", 0), 4))
]
log.setLevel(getattr(logging, verbosity.upper()))
formatter = logging.Formatter(
"%(asctime)s - %(name)s - %(levelname)s - %(message)s"
)
ch = logging.StreamHandler()
ch.setLevel(getattr(logging, verbosity.upper()))
ch.setFormatter(formatter)
log.addHandler(ch)
# GrapheneAPI logging
if ctx.obj.get("verbose", 0) > 4:
verbosity = ["critical", "error", "warn", "info", "debug"][
int(min(ctx.obj.get("verbose", 4) - 4, 4))
]
log = logging.getLogger("grapheneapi")
log.setLevel(getattr(logging, verbosity.upper()))
log.addHandler(ch)
if ctx.obj.get("verbose", 0) > 8:
verbosity = ["critical", "error", "warn", "info", "debug"][
int(min(ctx.obj.get("verbose", 8) - 8, 4))
]
log = logging.getLogger("graphenebase")
log.setLevel(getattr(logging, verbosity.upper()))
log.addHandler(ch)
return ctx.invoke(f, *args, **kwargs)
return update_wrapper(new_func, f) | [
"def",
"verbose",
"(",
"f",
")",
":",
"@",
"click",
".",
"pass_context",
"def",
"new_func",
"(",
"ctx",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"global",
"log",
"verbosity",
"=",
"[",
"\"critical\"",
",",
"\"error\"",
",",
"\"warn\"",
",",
"\"info\"",
",",
"\"debug\"",
"]",
"[",
"int",
"(",
"min",
"(",
"ctx",
".",
"obj",
".",
"get",
"(",
"\"verbose\"",
",",
"0",
")",
",",
"4",
")",
")",
"]",
"log",
".",
"setLevel",
"(",
"getattr",
"(",
"logging",
",",
"verbosity",
".",
"upper",
"(",
")",
")",
")",
"formatter",
"=",
"logging",
".",
"Formatter",
"(",
"\"%(asctime)s - %(name)s - %(levelname)s - %(message)s\"",
")",
"ch",
"=",
"logging",
".",
"StreamHandler",
"(",
")",
"ch",
".",
"setLevel",
"(",
"getattr",
"(",
"logging",
",",
"verbosity",
".",
"upper",
"(",
")",
")",
")",
"ch",
".",
"setFormatter",
"(",
"formatter",
")",
"log",
".",
"addHandler",
"(",
"ch",
")",
"# GrapheneAPI logging",
"if",
"ctx",
".",
"obj",
".",
"get",
"(",
"\"verbose\"",
",",
"0",
")",
">",
"4",
":",
"verbosity",
"=",
"[",
"\"critical\"",
",",
"\"error\"",
",",
"\"warn\"",
",",
"\"info\"",
",",
"\"debug\"",
"]",
"[",
"int",
"(",
"min",
"(",
"ctx",
".",
"obj",
".",
"get",
"(",
"\"verbose\"",
",",
"4",
")",
"-",
"4",
",",
"4",
")",
")",
"]",
"log",
"=",
"logging",
".",
"getLogger",
"(",
"\"grapheneapi\"",
")",
"log",
".",
"setLevel",
"(",
"getattr",
"(",
"logging",
",",
"verbosity",
".",
"upper",
"(",
")",
")",
")",
"log",
".",
"addHandler",
"(",
"ch",
")",
"if",
"ctx",
".",
"obj",
".",
"get",
"(",
"\"verbose\"",
",",
"0",
")",
">",
"8",
":",
"verbosity",
"=",
"[",
"\"critical\"",
",",
"\"error\"",
",",
"\"warn\"",
",",
"\"info\"",
",",
"\"debug\"",
"]",
"[",
"int",
"(",
"min",
"(",
"ctx",
".",
"obj",
".",
"get",
"(",
"\"verbose\"",
",",
"8",
")",
"-",
"8",
",",
"4",
")",
")",
"]",
"log",
"=",
"logging",
".",
"getLogger",
"(",
"\"graphenebase\"",
")",
"log",
".",
"setLevel",
"(",
"getattr",
"(",
"logging",
",",
"verbosity",
".",
"upper",
"(",
")",
")",
")",
"log",
".",
"addHandler",
"(",
"ch",
")",
"return",
"ctx",
".",
"invoke",
"(",
"f",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"return",
"update_wrapper",
"(",
"new_func",
",",
"f",
")"
] | Add verbose flags and add logging handlers | [
"Add",
"verbose",
"flags",
"and",
"add",
"logging",
"handlers"
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplays/cli/decorators.py#L13-L51 | train |
peerplays-network/python-peerplays | peerplays/cli/decorators.py | offline | def offline(f):
""" This decorator allows you to access ``ctx.peerplays`` which is
an instance of PeerPlays with ``offline=True``.
"""
@click.pass_context
@verbose
def new_func(ctx, *args, **kwargs):
ctx.obj["offline"] = True
ctx.peerplays = PeerPlays(**ctx.obj)
ctx.blockchain = ctx.peerplays
set_shared_peerplays_instance(ctx.peerplays)
return ctx.invoke(f, *args, **kwargs)
return update_wrapper(new_func, f) | python | def offline(f):
""" This decorator allows you to access ``ctx.peerplays`` which is
an instance of PeerPlays with ``offline=True``.
"""
@click.pass_context
@verbose
def new_func(ctx, *args, **kwargs):
ctx.obj["offline"] = True
ctx.peerplays = PeerPlays(**ctx.obj)
ctx.blockchain = ctx.peerplays
set_shared_peerplays_instance(ctx.peerplays)
return ctx.invoke(f, *args, **kwargs)
return update_wrapper(new_func, f) | [
"def",
"offline",
"(",
"f",
")",
":",
"@",
"click",
".",
"pass_context",
"@",
"verbose",
"def",
"new_func",
"(",
"ctx",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
".",
"obj",
"[",
"\"offline\"",
"]",
"=",
"True",
"ctx",
".",
"peerplays",
"=",
"PeerPlays",
"(",
"*",
"*",
"ctx",
".",
"obj",
")",
"ctx",
".",
"blockchain",
"=",
"ctx",
".",
"peerplays",
"set_shared_peerplays_instance",
"(",
"ctx",
".",
"peerplays",
")",
"return",
"ctx",
".",
"invoke",
"(",
"f",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"return",
"update_wrapper",
"(",
"new_func",
",",
"f",
")"
] | This decorator allows you to access ``ctx.peerplays`` which is
an instance of PeerPlays with ``offline=True``. | [
"This",
"decorator",
"allows",
"you",
"to",
"access",
"ctx",
".",
"peerplays",
"which",
"is",
"an",
"instance",
"of",
"PeerPlays",
"with",
"offline",
"=",
"True",
"."
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplays/cli/decorators.py#L54-L68 | train |
peerplays-network/python-peerplays | peerplays/cli/decorators.py | configfile | def configfile(f):
""" This decorator will parse a configuration file in YAML format
and store the dictionary in ``ctx.blockchain.config``
"""
@click.pass_context
def new_func(ctx, *args, **kwargs):
ctx.config = yaml.load(open(ctx.obj["configfile"]))
return ctx.invoke(f, *args, **kwargs)
return update_wrapper(new_func, f) | python | def configfile(f):
""" This decorator will parse a configuration file in YAML format
and store the dictionary in ``ctx.blockchain.config``
"""
@click.pass_context
def new_func(ctx, *args, **kwargs):
ctx.config = yaml.load(open(ctx.obj["configfile"]))
return ctx.invoke(f, *args, **kwargs)
return update_wrapper(new_func, f) | [
"def",
"configfile",
"(",
"f",
")",
":",
"@",
"click",
".",
"pass_context",
"def",
"new_func",
"(",
"ctx",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"ctx",
".",
"config",
"=",
"yaml",
".",
"load",
"(",
"open",
"(",
"ctx",
".",
"obj",
"[",
"\"configfile\"",
"]",
")",
")",
"return",
"ctx",
".",
"invoke",
"(",
"f",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"return",
"update_wrapper",
"(",
"new_func",
",",
"f",
")"
] | This decorator will parse a configuration file in YAML format
and store the dictionary in ``ctx.blockchain.config`` | [
"This",
"decorator",
"will",
"parse",
"a",
"configuration",
"file",
"in",
"YAML",
"format",
"and",
"store",
"the",
"dictionary",
"in",
"ctx",
".",
"blockchain",
".",
"config"
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplays/cli/decorators.py#L154-L164 | train |
peerplays-network/python-peerplays | peerplaysapi/websocket.py | PeerPlaysWebsocket.on_message | def on_message(self, ws, reply, *args):
""" This method is called by the websocket connection on every
message that is received. If we receive a ``notice``, we
hand over post-processing and signalling of events to
``process_notice``.
"""
log.debug("Received message: %s" % str(reply))
data = {}
try:
data = json.loads(reply, strict=False)
except ValueError:
raise ValueError("API node returned invalid format. Expected JSON!")
if data.get("method") == "notice":
id = data["params"][0]
if id >= len(self.__events__):
log.critical("Received an id that is out of range\n\n" + str(data))
return
# This is a "general" object change notification
if id == self.__events__.index("on_object"):
# Let's see if a specific object has changed
for notice in data["params"][1]:
try:
if "id" in notice:
self.process_notice(notice)
else:
for obj in notice:
if "id" in obj:
self.process_notice(obj)
except Exception as e:
log.critical(
"Error in process_notice: {}\n\n{}".format(
str(e), traceback.format_exc
)
)
else:
try:
callbackname = self.__events__[id]
log.info("Patching through to call %s" % callbackname)
[getattr(self.events, callbackname)(x) for x in data["params"][1]]
except Exception as e:
log.critical(
"Error in {}: {}\n\n{}".format(
callbackname, str(e), traceback.format_exc()
)
) | python | def on_message(self, ws, reply, *args):
""" This method is called by the websocket connection on every
message that is received. If we receive a ``notice``, we
hand over post-processing and signalling of events to
``process_notice``.
"""
log.debug("Received message: %s" % str(reply))
data = {}
try:
data = json.loads(reply, strict=False)
except ValueError:
raise ValueError("API node returned invalid format. Expected JSON!")
if data.get("method") == "notice":
id = data["params"][0]
if id >= len(self.__events__):
log.critical("Received an id that is out of range\n\n" + str(data))
return
# This is a "general" object change notification
if id == self.__events__.index("on_object"):
# Let's see if a specific object has changed
for notice in data["params"][1]:
try:
if "id" in notice:
self.process_notice(notice)
else:
for obj in notice:
if "id" in obj:
self.process_notice(obj)
except Exception as e:
log.critical(
"Error in process_notice: {}\n\n{}".format(
str(e), traceback.format_exc
)
)
else:
try:
callbackname = self.__events__[id]
log.info("Patching through to call %s" % callbackname)
[getattr(self.events, callbackname)(x) for x in data["params"][1]]
except Exception as e:
log.critical(
"Error in {}: {}\n\n{}".format(
callbackname, str(e), traceback.format_exc()
)
) | [
"def",
"on_message",
"(",
"self",
",",
"ws",
",",
"reply",
",",
"*",
"args",
")",
":",
"log",
".",
"debug",
"(",
"\"Received message: %s\"",
"%",
"str",
"(",
"reply",
")",
")",
"data",
"=",
"{",
"}",
"try",
":",
"data",
"=",
"json",
".",
"loads",
"(",
"reply",
",",
"strict",
"=",
"False",
")",
"except",
"ValueError",
":",
"raise",
"ValueError",
"(",
"\"API node returned invalid format. Expected JSON!\"",
")",
"if",
"data",
".",
"get",
"(",
"\"method\"",
")",
"==",
"\"notice\"",
":",
"id",
"=",
"data",
"[",
"\"params\"",
"]",
"[",
"0",
"]",
"if",
"id",
">=",
"len",
"(",
"self",
".",
"__events__",
")",
":",
"log",
".",
"critical",
"(",
"\"Received an id that is out of range\\n\\n\"",
"+",
"str",
"(",
"data",
")",
")",
"return",
"# This is a \"general\" object change notification",
"if",
"id",
"==",
"self",
".",
"__events__",
".",
"index",
"(",
"\"on_object\"",
")",
":",
"# Let's see if a specific object has changed",
"for",
"notice",
"in",
"data",
"[",
"\"params\"",
"]",
"[",
"1",
"]",
":",
"try",
":",
"if",
"\"id\"",
"in",
"notice",
":",
"self",
".",
"process_notice",
"(",
"notice",
")",
"else",
":",
"for",
"obj",
"in",
"notice",
":",
"if",
"\"id\"",
"in",
"obj",
":",
"self",
".",
"process_notice",
"(",
"obj",
")",
"except",
"Exception",
"as",
"e",
":",
"log",
".",
"critical",
"(",
"\"Error in process_notice: {}\\n\\n{}\"",
".",
"format",
"(",
"str",
"(",
"e",
")",
",",
"traceback",
".",
"format_exc",
")",
")",
"else",
":",
"try",
":",
"callbackname",
"=",
"self",
".",
"__events__",
"[",
"id",
"]",
"log",
".",
"info",
"(",
"\"Patching through to call %s\"",
"%",
"callbackname",
")",
"[",
"getattr",
"(",
"self",
".",
"events",
",",
"callbackname",
")",
"(",
"x",
")",
"for",
"x",
"in",
"data",
"[",
"\"params\"",
"]",
"[",
"1",
"]",
"]",
"except",
"Exception",
"as",
"e",
":",
"log",
".",
"critical",
"(",
"\"Error in {}: {}\\n\\n{}\"",
".",
"format",
"(",
"callbackname",
",",
"str",
"(",
"e",
")",
",",
"traceback",
".",
"format_exc",
"(",
")",
")",
")"
] | This method is called by the websocket connection on every
message that is received. If we receive a ``notice``, we
hand over post-processing and signalling of events to
``process_notice``. | [
"This",
"method",
"is",
"called",
"by",
"the",
"websocket",
"connection",
"on",
"every",
"message",
"that",
"is",
"received",
".",
"If",
"we",
"receive",
"a",
"notice",
"we",
"hand",
"over",
"post",
"-",
"processing",
"and",
"signalling",
"of",
"events",
"to",
"process_notice",
"."
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplaysapi/websocket.py#L216-L263 | train |
peerplays-network/python-peerplays | peerplaysapi/websocket.py | PeerPlaysWebsocket.on_close | def on_close(self, ws):
""" Called when websocket connection is closed
"""
log.debug("Closing WebSocket connection with {}".format(self.url))
if self.keepalive and self.keepalive.is_alive():
self.keepalive.do_run = False
self.keepalive.join() | python | def on_close(self, ws):
""" Called when websocket connection is closed
"""
log.debug("Closing WebSocket connection with {}".format(self.url))
if self.keepalive and self.keepalive.is_alive():
self.keepalive.do_run = False
self.keepalive.join() | [
"def",
"on_close",
"(",
"self",
",",
"ws",
")",
":",
"log",
".",
"debug",
"(",
"\"Closing WebSocket connection with {}\"",
".",
"format",
"(",
"self",
".",
"url",
")",
")",
"if",
"self",
".",
"keepalive",
"and",
"self",
".",
"keepalive",
".",
"is_alive",
"(",
")",
":",
"self",
".",
"keepalive",
".",
"do_run",
"=",
"False",
"self",
".",
"keepalive",
".",
"join",
"(",
")"
] | Called when websocket connection is closed | [
"Called",
"when",
"websocket",
"connection",
"is",
"closed"
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplaysapi/websocket.py#L270-L276 | train |
peerplays-network/python-peerplays | peerplaysapi/websocket.py | PeerPlaysWebsocket.run_forever | def run_forever(self):
""" This method is used to run the websocket app continuously.
It will execute callbacks as defined and try to stay
connected with the provided APIs
"""
cnt = 0
while True:
cnt += 1
self.url = next(self.urls)
log.debug("Trying to connect to node %s" % self.url)
try:
# websocket.enableTrace(True)
self.ws = websocket.WebSocketApp(
self.url,
on_message=self.on_message,
# on_data=self.on_message,
on_error=self.on_error,
on_close=self.on_close,
on_open=self.on_open,
)
self.ws.run_forever()
except websocket.WebSocketException as exc:
if self.num_retries >= 0 and cnt > self.num_retries:
raise NumRetriesReached()
sleeptime = (cnt - 1) * 2 if cnt < 10 else 10
if sleeptime:
log.warning(
"Lost connection to node during wsconnect(): %s (%d/%d) "
% (self.url, cnt, self.num_retries)
+ "Retrying in %d seconds" % sleeptime
)
time.sleep(sleeptime)
except KeyboardInterrupt:
self.ws.keep_running = False
raise
except Exception as e:
log.critical("{}\n\n{}".format(str(e), traceback.format_exc())) | python | def run_forever(self):
""" This method is used to run the websocket app continuously.
It will execute callbacks as defined and try to stay
connected with the provided APIs
"""
cnt = 0
while True:
cnt += 1
self.url = next(self.urls)
log.debug("Trying to connect to node %s" % self.url)
try:
# websocket.enableTrace(True)
self.ws = websocket.WebSocketApp(
self.url,
on_message=self.on_message,
# on_data=self.on_message,
on_error=self.on_error,
on_close=self.on_close,
on_open=self.on_open,
)
self.ws.run_forever()
except websocket.WebSocketException as exc:
if self.num_retries >= 0 and cnt > self.num_retries:
raise NumRetriesReached()
sleeptime = (cnt - 1) * 2 if cnt < 10 else 10
if sleeptime:
log.warning(
"Lost connection to node during wsconnect(): %s (%d/%d) "
% (self.url, cnt, self.num_retries)
+ "Retrying in %d seconds" % sleeptime
)
time.sleep(sleeptime)
except KeyboardInterrupt:
self.ws.keep_running = False
raise
except Exception as e:
log.critical("{}\n\n{}".format(str(e), traceback.format_exc())) | [
"def",
"run_forever",
"(",
"self",
")",
":",
"cnt",
"=",
"0",
"while",
"True",
":",
"cnt",
"+=",
"1",
"self",
".",
"url",
"=",
"next",
"(",
"self",
".",
"urls",
")",
"log",
".",
"debug",
"(",
"\"Trying to connect to node %s\"",
"%",
"self",
".",
"url",
")",
"try",
":",
"# websocket.enableTrace(True)",
"self",
".",
"ws",
"=",
"websocket",
".",
"WebSocketApp",
"(",
"self",
".",
"url",
",",
"on_message",
"=",
"self",
".",
"on_message",
",",
"# on_data=self.on_message,",
"on_error",
"=",
"self",
".",
"on_error",
",",
"on_close",
"=",
"self",
".",
"on_close",
",",
"on_open",
"=",
"self",
".",
"on_open",
",",
")",
"self",
".",
"ws",
".",
"run_forever",
"(",
")",
"except",
"websocket",
".",
"WebSocketException",
"as",
"exc",
":",
"if",
"self",
".",
"num_retries",
">=",
"0",
"and",
"cnt",
">",
"self",
".",
"num_retries",
":",
"raise",
"NumRetriesReached",
"(",
")",
"sleeptime",
"=",
"(",
"cnt",
"-",
"1",
")",
"*",
"2",
"if",
"cnt",
"<",
"10",
"else",
"10",
"if",
"sleeptime",
":",
"log",
".",
"warning",
"(",
"\"Lost connection to node during wsconnect(): %s (%d/%d) \"",
"%",
"(",
"self",
".",
"url",
",",
"cnt",
",",
"self",
".",
"num_retries",
")",
"+",
"\"Retrying in %d seconds\"",
"%",
"sleeptime",
")",
"time",
".",
"sleep",
"(",
"sleeptime",
")",
"except",
"KeyboardInterrupt",
":",
"self",
".",
"ws",
".",
"keep_running",
"=",
"False",
"raise",
"except",
"Exception",
"as",
"e",
":",
"log",
".",
"critical",
"(",
"\"{}\\n\\n{}\"",
".",
"format",
"(",
"str",
"(",
"e",
")",
",",
"traceback",
".",
"format_exc",
"(",
")",
")",
")"
] | This method is used to run the websocket app continuously.
It will execute callbacks as defined and try to stay
connected with the provided APIs | [
"This",
"method",
"is",
"used",
"to",
"run",
"the",
"websocket",
"app",
"continuously",
".",
"It",
"will",
"execute",
"callbacks",
"as",
"defined",
"and",
"try",
"to",
"stay",
"connected",
"with",
"the",
"provided",
"APIs"
] | 188f04238e7e21d5f73e9b01099eea44289ef6b7 | https://github.com/peerplays-network/python-peerplays/blob/188f04238e7e21d5f73e9b01099eea44289ef6b7/peerplaysapi/websocket.py#L278-L317 | train |
Zsailer/pandas_flavor | pandas_flavor/register.py | register_dataframe_method | def register_dataframe_method(method):
"""Register a function as a method attached to the Pandas DataFrame.
Example
-------
.. code-block:: python
@register_dataframe_method
def print_column(df, col):
'''Print the dataframe column given'''
print(df[col])
"""
def inner(*args, **kwargs):
class AccessorMethod(object):
def __init__(self, pandas_obj):
self._obj = pandas_obj
@wraps(method)
def __call__(self, *args, **kwargs):
return method(self._obj, *args, **kwargs)
register_dataframe_accessor(method.__name__)(AccessorMethod)
return method
return inner() | python | def register_dataframe_method(method):
"""Register a function as a method attached to the Pandas DataFrame.
Example
-------
.. code-block:: python
@register_dataframe_method
def print_column(df, col):
'''Print the dataframe column given'''
print(df[col])
"""
def inner(*args, **kwargs):
class AccessorMethod(object):
def __init__(self, pandas_obj):
self._obj = pandas_obj
@wraps(method)
def __call__(self, *args, **kwargs):
return method(self._obj, *args, **kwargs)
register_dataframe_accessor(method.__name__)(AccessorMethod)
return method
return inner() | [
"def",
"register_dataframe_method",
"(",
"method",
")",
":",
"def",
"inner",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"class",
"AccessorMethod",
"(",
"object",
")",
":",
"def",
"__init__",
"(",
"self",
",",
"pandas_obj",
")",
":",
"self",
".",
"_obj",
"=",
"pandas_obj",
"@",
"wraps",
"(",
"method",
")",
"def",
"__call__",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"method",
"(",
"self",
".",
"_obj",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"register_dataframe_accessor",
"(",
"method",
".",
"__name__",
")",
"(",
"AccessorMethod",
")",
"return",
"method",
"return",
"inner",
"(",
")"
] | Register a function as a method attached to the Pandas DataFrame.
Example
-------
.. code-block:: python
@register_dataframe_method
def print_column(df, col):
'''Print the dataframe column given'''
print(df[col]) | [
"Register",
"a",
"function",
"as",
"a",
"method",
"attached",
"to",
"the",
"Pandas",
"DataFrame",
"."
] | 1953aeee09424300d69a11dd2ffd3460a806fb65 | https://github.com/Zsailer/pandas_flavor/blob/1953aeee09424300d69a11dd2ffd3460a806fb65/pandas_flavor/register.py#L6-L35 | train |
Zsailer/pandas_flavor | pandas_flavor/register.py | register_series_method | def register_series_method(method):
"""Register a function as a method attached to the Pandas Series.
"""
def inner(*args, **kwargs):
class AccessorMethod(object):
__doc__ = method.__doc__
def __init__(self, pandas_obj):
self._obj = pandas_obj
@wraps(method)
def __call__(self, *args, **kwargs):
return method(self._obj, *args, **kwargs)
register_series_accessor(method.__name__)(AccessorMethod)
return method
return inner() | python | def register_series_method(method):
"""Register a function as a method attached to the Pandas Series.
"""
def inner(*args, **kwargs):
class AccessorMethod(object):
__doc__ = method.__doc__
def __init__(self, pandas_obj):
self._obj = pandas_obj
@wraps(method)
def __call__(self, *args, **kwargs):
return method(self._obj, *args, **kwargs)
register_series_accessor(method.__name__)(AccessorMethod)
return method
return inner() | [
"def",
"register_series_method",
"(",
"method",
")",
":",
"def",
"inner",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"class",
"AccessorMethod",
"(",
"object",
")",
":",
"__doc__",
"=",
"method",
".",
"__doc__",
"def",
"__init__",
"(",
"self",
",",
"pandas_obj",
")",
":",
"self",
".",
"_obj",
"=",
"pandas_obj",
"@",
"wraps",
"(",
"method",
")",
"def",
"__call__",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"return",
"method",
"(",
"self",
".",
"_obj",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"register_series_accessor",
"(",
"method",
".",
"__name__",
")",
"(",
"AccessorMethod",
")",
"return",
"method",
"return",
"inner",
"(",
")"
] | Register a function as a method attached to the Pandas Series. | [
"Register",
"a",
"function",
"as",
"a",
"method",
"attached",
"to",
"the",
"Pandas",
"Series",
"."
] | 1953aeee09424300d69a11dd2ffd3460a806fb65 | https://github.com/Zsailer/pandas_flavor/blob/1953aeee09424300d69a11dd2ffd3460a806fb65/pandas_flavor/register.py#L38-L57 | train |
pinax/pinax-invitations | pinax/invitations/models.py | InvitationStat.add_invites_to_user | def add_invites_to_user(cls, user, amount):
"""
Add the specified number of invites to current allocated total.
"""
stat, _ = InvitationStat.objects.get_or_create(user=user)
if stat.invites_allocated != -1:
stat.invites_allocated += amount
stat.save() | python | def add_invites_to_user(cls, user, amount):
"""
Add the specified number of invites to current allocated total.
"""
stat, _ = InvitationStat.objects.get_or_create(user=user)
if stat.invites_allocated != -1:
stat.invites_allocated += amount
stat.save() | [
"def",
"add_invites_to_user",
"(",
"cls",
",",
"user",
",",
"amount",
")",
":",
"stat",
",",
"_",
"=",
"InvitationStat",
".",
"objects",
".",
"get_or_create",
"(",
"user",
"=",
"user",
")",
"if",
"stat",
".",
"invites_allocated",
"!=",
"-",
"1",
":",
"stat",
".",
"invites_allocated",
"+=",
"amount",
"stat",
".",
"save",
"(",
")"
] | Add the specified number of invites to current allocated total. | [
"Add",
"the",
"specified",
"number",
"of",
"invites",
"to",
"current",
"allocated",
"total",
"."
] | 6c6e863da179a1c620074efe5b5728cd1e6eff1b | https://github.com/pinax/pinax-invitations/blob/6c6e863da179a1c620074efe5b5728cd1e6eff1b/pinax/invitations/models.py#L111-L118 | train |
pinax/pinax-invitations | pinax/invitations/models.py | InvitationStat.add_invites | def add_invites(cls, amount):
"""
Add invites for all users.
"""
for user in get_user_model().objects.all():
cls.add_invites_to_user(user, amount) | python | def add_invites(cls, amount):
"""
Add invites for all users.
"""
for user in get_user_model().objects.all():
cls.add_invites_to_user(user, amount) | [
"def",
"add_invites",
"(",
"cls",
",",
"amount",
")",
":",
"for",
"user",
"in",
"get_user_model",
"(",
")",
".",
"objects",
".",
"all",
"(",
")",
":",
"cls",
".",
"add_invites_to_user",
"(",
"user",
",",
"amount",
")"
] | Add invites for all users. | [
"Add",
"invites",
"for",
"all",
"users",
"."
] | 6c6e863da179a1c620074efe5b5728cd1e6eff1b | https://github.com/pinax/pinax-invitations/blob/6c6e863da179a1c620074efe5b5728cd1e6eff1b/pinax/invitations/models.py#L121-L126 | train |
pinax/pinax-invitations | pinax/invitations/models.py | InvitationStat.topoff_user | def topoff_user(cls, user, amount):
"""
Ensure user has a minimum number of invites.
"""
stat, _ = cls.objects.get_or_create(user=user)
remaining = stat.invites_remaining()
if remaining != -1 and remaining < amount:
stat.invites_allocated += (amount - remaining)
stat.save() | python | def topoff_user(cls, user, amount):
"""
Ensure user has a minimum number of invites.
"""
stat, _ = cls.objects.get_or_create(user=user)
remaining = stat.invites_remaining()
if remaining != -1 and remaining < amount:
stat.invites_allocated += (amount - remaining)
stat.save() | [
"def",
"topoff_user",
"(",
"cls",
",",
"user",
",",
"amount",
")",
":",
"stat",
",",
"_",
"=",
"cls",
".",
"objects",
".",
"get_or_create",
"(",
"user",
"=",
"user",
")",
"remaining",
"=",
"stat",
".",
"invites_remaining",
"(",
")",
"if",
"remaining",
"!=",
"-",
"1",
"and",
"remaining",
"<",
"amount",
":",
"stat",
".",
"invites_allocated",
"+=",
"(",
"amount",
"-",
"remaining",
")",
"stat",
".",
"save",
"(",
")"
] | Ensure user has a minimum number of invites. | [
"Ensure",
"user",
"has",
"a",
"minimum",
"number",
"of",
"invites",
"."
] | 6c6e863da179a1c620074efe5b5728cd1e6eff1b | https://github.com/pinax/pinax-invitations/blob/6c6e863da179a1c620074efe5b5728cd1e6eff1b/pinax/invitations/models.py#L129-L137 | train |
pinax/pinax-invitations | pinax/invitations/models.py | InvitationStat.topoff | def topoff(cls, amount):
"""
Ensure all users have a minimum number of invites.
"""
for user in get_user_model().objects.all():
cls.topoff_user(user, amount) | python | def topoff(cls, amount):
"""
Ensure all users have a minimum number of invites.
"""
for user in get_user_model().objects.all():
cls.topoff_user(user, amount) | [
"def",
"topoff",
"(",
"cls",
",",
"amount",
")",
":",
"for",
"user",
"in",
"get_user_model",
"(",
")",
".",
"objects",
".",
"all",
"(",
")",
":",
"cls",
".",
"topoff_user",
"(",
"user",
",",
"amount",
")"
] | Ensure all users have a minimum number of invites. | [
"Ensure",
"all",
"users",
"have",
"a",
"minimum",
"number",
"of",
"invites",
"."
] | 6c6e863da179a1c620074efe5b5728cd1e6eff1b | https://github.com/pinax/pinax-invitations/blob/6c6e863da179a1c620074efe5b5728cd1e6eff1b/pinax/invitations/models.py#L140-L145 | train |
skelsec/minidump | minidump/minidumpreader.py | MinidumpBufferedReader.align | def align(self, alignment = None):
"""
Repositions the current reader to match architecture alignment
"""
if alignment is None:
if self.reader.sysinfo.ProcessorArchitecture == PROCESSOR_ARCHITECTURE.AMD64:
alignment = 8
else:
alignment = 4
offset = self.current_position % alignment
if offset == 0:
return
offset_to_aligned = (alignment - offset) % alignment
self.seek(offset_to_aligned, 1)
return | python | def align(self, alignment = None):
"""
Repositions the current reader to match architecture alignment
"""
if alignment is None:
if self.reader.sysinfo.ProcessorArchitecture == PROCESSOR_ARCHITECTURE.AMD64:
alignment = 8
else:
alignment = 4
offset = self.current_position % alignment
if offset == 0:
return
offset_to_aligned = (alignment - offset) % alignment
self.seek(offset_to_aligned, 1)
return | [
"def",
"align",
"(",
"self",
",",
"alignment",
"=",
"None",
")",
":",
"if",
"alignment",
"is",
"None",
":",
"if",
"self",
".",
"reader",
".",
"sysinfo",
".",
"ProcessorArchitecture",
"==",
"PROCESSOR_ARCHITECTURE",
".",
"AMD64",
":",
"alignment",
"=",
"8",
"else",
":",
"alignment",
"=",
"4",
"offset",
"=",
"self",
".",
"current_position",
"%",
"alignment",
"if",
"offset",
"==",
"0",
":",
"return",
"offset_to_aligned",
"=",
"(",
"alignment",
"-",
"offset",
")",
"%",
"alignment",
"self",
".",
"seek",
"(",
"offset_to_aligned",
",",
"1",
")",
"return"
] | Repositions the current reader to match architecture alignment | [
"Repositions",
"the",
"current",
"reader",
"to",
"match",
"architecture",
"alignment"
] | 0c4dcabe6f11d7a403440919ffa9e3c9889c5212 | https://github.com/skelsec/minidump/blob/0c4dcabe6f11d7a403440919ffa9e3c9889c5212/minidump/minidumpreader.py#L87-L101 | train |
skelsec/minidump | minidump/minidumpreader.py | MinidumpBufferedReader.peek | def peek(self, length):
"""
Returns up to length bytes from the current memory segment
"""
t = self.current_position + length
if not self.current_segment.inrange(t):
raise Exception('Would read over segment boundaries!')
return self.current_segment.data[self.current_position - self.current_segment.start_address :t - self.current_segment.start_address] | python | def peek(self, length):
"""
Returns up to length bytes from the current memory segment
"""
t = self.current_position + length
if not self.current_segment.inrange(t):
raise Exception('Would read over segment boundaries!')
return self.current_segment.data[self.current_position - self.current_segment.start_address :t - self.current_segment.start_address] | [
"def",
"peek",
"(",
"self",
",",
"length",
")",
":",
"t",
"=",
"self",
".",
"current_position",
"+",
"length",
"if",
"not",
"self",
".",
"current_segment",
".",
"inrange",
"(",
"t",
")",
":",
"raise",
"Exception",
"(",
"'Would read over segment boundaries!'",
")",
"return",
"self",
".",
"current_segment",
".",
"data",
"[",
"self",
".",
"current_position",
"-",
"self",
".",
"current_segment",
".",
"start_address",
":",
"t",
"-",
"self",
".",
"current_segment",
".",
"start_address",
"]"
] | Returns up to length bytes from the current memory segment | [
"Returns",
"up",
"to",
"length",
"bytes",
"from",
"the",
"current",
"memory",
"segment"
] | 0c4dcabe6f11d7a403440919ffa9e3c9889c5212 | https://github.com/skelsec/minidump/blob/0c4dcabe6f11d7a403440919ffa9e3c9889c5212/minidump/minidumpreader.py#L109-L116 | train |
skelsec/minidump | minidump/minidumpreader.py | MinidumpBufferedReader.read | def read(self, size = -1):
"""
Returns data bytes of size size from the current segment. If size is -1 it returns all the remaining data bytes from memory segment
"""
if size < -1:
raise Exception('You shouldnt be doing this')
if size == -1:
t = self.current_segment.remaining_len(self.current_position)
if not t:
return None
old_new_pos = self.current_position
self.current_position = self.current_segment.end_address
return self.current_segment.data[old_new_pos - self.current_segment.start_address:]
t = self.current_position + size
if not self.current_segment.inrange(t):
raise Exception('Would read over segment boundaries!')
old_new_pos = self.current_position
self.current_position = t
return self.current_segment.data[old_new_pos - self.current_segment.start_address :t - self.current_segment.start_address] | python | def read(self, size = -1):
"""
Returns data bytes of size size from the current segment. If size is -1 it returns all the remaining data bytes from memory segment
"""
if size < -1:
raise Exception('You shouldnt be doing this')
if size == -1:
t = self.current_segment.remaining_len(self.current_position)
if not t:
return None
old_new_pos = self.current_position
self.current_position = self.current_segment.end_address
return self.current_segment.data[old_new_pos - self.current_segment.start_address:]
t = self.current_position + size
if not self.current_segment.inrange(t):
raise Exception('Would read over segment boundaries!')
old_new_pos = self.current_position
self.current_position = t
return self.current_segment.data[old_new_pos - self.current_segment.start_address :t - self.current_segment.start_address] | [
"def",
"read",
"(",
"self",
",",
"size",
"=",
"-",
"1",
")",
":",
"if",
"size",
"<",
"-",
"1",
":",
"raise",
"Exception",
"(",
"'You shouldnt be doing this'",
")",
"if",
"size",
"==",
"-",
"1",
":",
"t",
"=",
"self",
".",
"current_segment",
".",
"remaining_len",
"(",
"self",
".",
"current_position",
")",
"if",
"not",
"t",
":",
"return",
"None",
"old_new_pos",
"=",
"self",
".",
"current_position",
"self",
".",
"current_position",
"=",
"self",
".",
"current_segment",
".",
"end_address",
"return",
"self",
".",
"current_segment",
".",
"data",
"[",
"old_new_pos",
"-",
"self",
".",
"current_segment",
".",
"start_address",
":",
"]",
"t",
"=",
"self",
".",
"current_position",
"+",
"size",
"if",
"not",
"self",
".",
"current_segment",
".",
"inrange",
"(",
"t",
")",
":",
"raise",
"Exception",
"(",
"'Would read over segment boundaries!'",
")",
"old_new_pos",
"=",
"self",
".",
"current_position",
"self",
".",
"current_position",
"=",
"t",
"return",
"self",
".",
"current_segment",
".",
"data",
"[",
"old_new_pos",
"-",
"self",
".",
"current_segment",
".",
"start_address",
":",
"t",
"-",
"self",
".",
"current_segment",
".",
"start_address",
"]"
] | Returns data bytes of size size from the current segment. If size is -1 it returns all the remaining data bytes from memory segment | [
"Returns",
"data",
"bytes",
"of",
"size",
"size",
"from",
"the",
"current",
"segment",
".",
"If",
"size",
"is",
"-",
"1",
"it",
"returns",
"all",
"the",
"remaining",
"data",
"bytes",
"from",
"memory",
"segment"
] | 0c4dcabe6f11d7a403440919ffa9e3c9889c5212 | https://github.com/skelsec/minidump/blob/0c4dcabe6f11d7a403440919ffa9e3c9889c5212/minidump/minidumpreader.py#L118-L139 | train |
skelsec/minidump | minidump/minidumpreader.py | MinidumpBufferedReader.read_int | def read_int(self):
"""
Reads an integer. The size depends on the architecture.
Reads a 4 byte small-endian singed int on 32 bit arch
Reads an 8 byte small-endian singed int on 64 bit arch
"""
if self.reader.sysinfo.ProcessorArchitecture == PROCESSOR_ARCHITECTURE.AMD64:
return int.from_bytes(self.read(8), byteorder = 'little', signed = True)
else:
return int.from_bytes(self.read(4), byteorder = 'little', signed = True) | python | def read_int(self):
"""
Reads an integer. The size depends on the architecture.
Reads a 4 byte small-endian singed int on 32 bit arch
Reads an 8 byte small-endian singed int on 64 bit arch
"""
if self.reader.sysinfo.ProcessorArchitecture == PROCESSOR_ARCHITECTURE.AMD64:
return int.from_bytes(self.read(8), byteorder = 'little', signed = True)
else:
return int.from_bytes(self.read(4), byteorder = 'little', signed = True) | [
"def",
"read_int",
"(",
"self",
")",
":",
"if",
"self",
".",
"reader",
".",
"sysinfo",
".",
"ProcessorArchitecture",
"==",
"PROCESSOR_ARCHITECTURE",
".",
"AMD64",
":",
"return",
"int",
".",
"from_bytes",
"(",
"self",
".",
"read",
"(",
"8",
")",
",",
"byteorder",
"=",
"'little'",
",",
"signed",
"=",
"True",
")",
"else",
":",
"return",
"int",
".",
"from_bytes",
"(",
"self",
".",
"read",
"(",
"4",
")",
",",
"byteorder",
"=",
"'little'",
",",
"signed",
"=",
"True",
")"
] | Reads an integer. The size depends on the architecture.
Reads a 4 byte small-endian singed int on 32 bit arch
Reads an 8 byte small-endian singed int on 64 bit arch | [
"Reads",
"an",
"integer",
".",
"The",
"size",
"depends",
"on",
"the",
"architecture",
".",
"Reads",
"a",
"4",
"byte",
"small",
"-",
"endian",
"singed",
"int",
"on",
"32",
"bit",
"arch",
"Reads",
"an",
"8",
"byte",
"small",
"-",
"endian",
"singed",
"int",
"on",
"64",
"bit",
"arch"
] | 0c4dcabe6f11d7a403440919ffa9e3c9889c5212 | https://github.com/skelsec/minidump/blob/0c4dcabe6f11d7a403440919ffa9e3c9889c5212/minidump/minidumpreader.py#L141-L150 | train |
skelsec/minidump | minidump/minidumpreader.py | MinidumpBufferedReader.read_uint | def read_uint(self):
"""
Reads an integer. The size depends on the architecture.
Reads a 4 byte small-endian unsinged int on 32 bit arch
Reads an 8 byte small-endian unsinged int on 64 bit arch
"""
if self.reader.sysinfo.ProcessorArchitecture == PROCESSOR_ARCHITECTURE.AMD64:
return int.from_bytes(self.read(8), byteorder = 'little', signed = False)
else:
return int.from_bytes(self.read(4), byteorder = 'little', signed = False) | python | def read_uint(self):
"""
Reads an integer. The size depends on the architecture.
Reads a 4 byte small-endian unsinged int on 32 bit arch
Reads an 8 byte small-endian unsinged int on 64 bit arch
"""
if self.reader.sysinfo.ProcessorArchitecture == PROCESSOR_ARCHITECTURE.AMD64:
return int.from_bytes(self.read(8), byteorder = 'little', signed = False)
else:
return int.from_bytes(self.read(4), byteorder = 'little', signed = False) | [
"def",
"read_uint",
"(",
"self",
")",
":",
"if",
"self",
".",
"reader",
".",
"sysinfo",
".",
"ProcessorArchitecture",
"==",
"PROCESSOR_ARCHITECTURE",
".",
"AMD64",
":",
"return",
"int",
".",
"from_bytes",
"(",
"self",
".",
"read",
"(",
"8",
")",
",",
"byteorder",
"=",
"'little'",
",",
"signed",
"=",
"False",
")",
"else",
":",
"return",
"int",
".",
"from_bytes",
"(",
"self",
".",
"read",
"(",
"4",
")",
",",
"byteorder",
"=",
"'little'",
",",
"signed",
"=",
"False",
")"
] | Reads an integer. The size depends on the architecture.
Reads a 4 byte small-endian unsinged int on 32 bit arch
Reads an 8 byte small-endian unsinged int on 64 bit arch | [
"Reads",
"an",
"integer",
".",
"The",
"size",
"depends",
"on",
"the",
"architecture",
".",
"Reads",
"a",
"4",
"byte",
"small",
"-",
"endian",
"unsinged",
"int",
"on",
"32",
"bit",
"arch",
"Reads",
"an",
"8",
"byte",
"small",
"-",
"endian",
"unsinged",
"int",
"on",
"64",
"bit",
"arch"
] | 0c4dcabe6f11d7a403440919ffa9e3c9889c5212 | https://github.com/skelsec/minidump/blob/0c4dcabe6f11d7a403440919ffa9e3c9889c5212/minidump/minidumpreader.py#L152-L161 | train |
skelsec/minidump | minidump/minidumpreader.py | MinidumpBufferedReader.find | def find(self, pattern):
"""
Searches for a pattern in the current memory segment
"""
pos = self.current_segment.data.find(pattern)
if pos == -1:
return -1
return pos + self.current_position | python | def find(self, pattern):
"""
Searches for a pattern in the current memory segment
"""
pos = self.current_segment.data.find(pattern)
if pos == -1:
return -1
return pos + self.current_position | [
"def",
"find",
"(",
"self",
",",
"pattern",
")",
":",
"pos",
"=",
"self",
".",
"current_segment",
".",
"data",
".",
"find",
"(",
"pattern",
")",
"if",
"pos",
"==",
"-",
"1",
":",
"return",
"-",
"1",
"return",
"pos",
"+",
"self",
".",
"current_position"
] | Searches for a pattern in the current memory segment | [
"Searches",
"for",
"a",
"pattern",
"in",
"the",
"current",
"memory",
"segment"
] | 0c4dcabe6f11d7a403440919ffa9e3c9889c5212 | https://github.com/skelsec/minidump/blob/0c4dcabe6f11d7a403440919ffa9e3c9889c5212/minidump/minidumpreader.py#L163-L170 | train |
skelsec/minidump | minidump/minidumpreader.py | MinidumpBufferedReader.find_all | def find_all(self, pattern):
"""
Searches for all occurrences of a pattern in the current memory segment, returns all occurrences as a list
"""
pos = []
last_found = -1
while True:
last_found = self.current_segment.data.find(pattern, last_found + 1)
if last_found == -1:
break
pos.append(last_found + self.current_segment.start_address)
return pos | python | def find_all(self, pattern):
"""
Searches for all occurrences of a pattern in the current memory segment, returns all occurrences as a list
"""
pos = []
last_found = -1
while True:
last_found = self.current_segment.data.find(pattern, last_found + 1)
if last_found == -1:
break
pos.append(last_found + self.current_segment.start_address)
return pos | [
"def",
"find_all",
"(",
"self",
",",
"pattern",
")",
":",
"pos",
"=",
"[",
"]",
"last_found",
"=",
"-",
"1",
"while",
"True",
":",
"last_found",
"=",
"self",
".",
"current_segment",
".",
"data",
".",
"find",
"(",
"pattern",
",",
"last_found",
"+",
"1",
")",
"if",
"last_found",
"==",
"-",
"1",
":",
"break",
"pos",
".",
"append",
"(",
"last_found",
"+",
"self",
".",
"current_segment",
".",
"start_address",
")",
"return",
"pos"
] | Searches for all occurrences of a pattern in the current memory segment, returns all occurrences as a list | [
"Searches",
"for",
"all",
"occurrences",
"of",
"a",
"pattern",
"in",
"the",
"current",
"memory",
"segment",
"returns",
"all",
"occurrences",
"as",
"a",
"list"
] | 0c4dcabe6f11d7a403440919ffa9e3c9889c5212 | https://github.com/skelsec/minidump/blob/0c4dcabe6f11d7a403440919ffa9e3c9889c5212/minidump/minidumpreader.py#L172-L184 | train |
skelsec/minidump | minidump/minidumpreader.py | MinidumpBufferedReader.find_global | def find_global(self, pattern):
"""
Searches for the pattern in the whole process memory space and returns the first occurrence.
This is exhaustive!
"""
pos_s = self.reader.search(pattern)
if len(pos_s) == 0:
return -1
return pos_s[0] | python | def find_global(self, pattern):
"""
Searches for the pattern in the whole process memory space and returns the first occurrence.
This is exhaustive!
"""
pos_s = self.reader.search(pattern)
if len(pos_s) == 0:
return -1
return pos_s[0] | [
"def",
"find_global",
"(",
"self",
",",
"pattern",
")",
":",
"pos_s",
"=",
"self",
".",
"reader",
".",
"search",
"(",
"pattern",
")",
"if",
"len",
"(",
"pos_s",
")",
"==",
"0",
":",
"return",
"-",
"1",
"return",
"pos_s",
"[",
"0",
"]"
] | Searches for the pattern in the whole process memory space and returns the first occurrence.
This is exhaustive! | [
"Searches",
"for",
"the",
"pattern",
"in",
"the",
"whole",
"process",
"memory",
"space",
"and",
"returns",
"the",
"first",
"occurrence",
".",
"This",
"is",
"exhaustive!"
] | 0c4dcabe6f11d7a403440919ffa9e3c9889c5212 | https://github.com/skelsec/minidump/blob/0c4dcabe6f11d7a403440919ffa9e3c9889c5212/minidump/minidumpreader.py#L186-L195 | train |
skelsec/minidump | minidump/utils/privileges.py | report_privilege_information | def report_privilege_information():
"Report all privilege information assigned to the current process."
privileges = get_privilege_information()
print("found {0} privileges".format(privileges.count))
tuple(map(print, privileges)) | python | def report_privilege_information():
"Report all privilege information assigned to the current process."
privileges = get_privilege_information()
print("found {0} privileges".format(privileges.count))
tuple(map(print, privileges)) | [
"def",
"report_privilege_information",
"(",
")",
":",
"privileges",
"=",
"get_privilege_information",
"(",
")",
"print",
"(",
"\"found {0} privileges\"",
".",
"format",
"(",
"privileges",
".",
"count",
")",
")",
"tuple",
"(",
"map",
"(",
"print",
",",
"privileges",
")",
")"
] | Report all privilege information assigned to the current process. | [
"Report",
"all",
"privilege",
"information",
"assigned",
"to",
"the",
"current",
"process",
"."
] | 0c4dcabe6f11d7a403440919ffa9e3c9889c5212 | https://github.com/skelsec/minidump/blob/0c4dcabe6f11d7a403440919ffa9e3c9889c5212/minidump/utils/privileges.py#L171-L175 | train |
rajasimon/beatserver | beatserver/server.py | BeatServer.handle | async def handle(self):
"""
Listens on all the provided channels and handles the messages.
"""
# For each channel, launch its own listening coroutine
listeners = []
for key, value in self.beat_config.items():
listeners.append(asyncio.ensure_future(
self.listener(key)
))
# For each beat configuration, launch it's own sending pattern
emitters = []
for key, value in self.beat_config.items():
emitters.append(asyncio.ensure_future(
self.emitters(key, value)
))
# Wait for them all to exit
await asyncio.wait(emitters)
await asyncio.wait(listeners) | python | async def handle(self):
"""
Listens on all the provided channels and handles the messages.
"""
# For each channel, launch its own listening coroutine
listeners = []
for key, value in self.beat_config.items():
listeners.append(asyncio.ensure_future(
self.listener(key)
))
# For each beat configuration, launch it's own sending pattern
emitters = []
for key, value in self.beat_config.items():
emitters.append(asyncio.ensure_future(
self.emitters(key, value)
))
# Wait for them all to exit
await asyncio.wait(emitters)
await asyncio.wait(listeners) | [
"async",
"def",
"handle",
"(",
"self",
")",
":",
"# For each channel, launch its own listening coroutine",
"listeners",
"=",
"[",
"]",
"for",
"key",
",",
"value",
"in",
"self",
".",
"beat_config",
".",
"items",
"(",
")",
":",
"listeners",
".",
"append",
"(",
"asyncio",
".",
"ensure_future",
"(",
"self",
".",
"listener",
"(",
"key",
")",
")",
")",
"# For each beat configuration, launch it's own sending pattern",
"emitters",
"=",
"[",
"]",
"for",
"key",
",",
"value",
"in",
"self",
".",
"beat_config",
".",
"items",
"(",
")",
":",
"emitters",
".",
"append",
"(",
"asyncio",
".",
"ensure_future",
"(",
"self",
".",
"emitters",
"(",
"key",
",",
"value",
")",
")",
")",
"# Wait for them all to exit",
"await",
"asyncio",
".",
"wait",
"(",
"emitters",
")",
"await",
"asyncio",
".",
"wait",
"(",
"listeners",
")"
] | Listens on all the provided channels and handles the messages. | [
"Listens",
"on",
"all",
"the",
"provided",
"channels",
"and",
"handles",
"the",
"messages",
"."
] | 8c653c46cdcf98398ca9d0bc6d3e47e5d621bb6a | https://github.com/rajasimon/beatserver/blob/8c653c46cdcf98398ca9d0bc6d3e47e5d621bb6a/beatserver/server.py#L16-L37 | train |
rajasimon/beatserver | beatserver/server.py | BeatServer.emitters | async def emitters(self, key, value):
"""
Single-channel emitter
"""
while True:
await asyncio.sleep(value['schedule'].total_seconds())
await self.channel_layer.send(key, {
"type": value['type'],
"message": value['message']
}) | python | async def emitters(self, key, value):
"""
Single-channel emitter
"""
while True:
await asyncio.sleep(value['schedule'].total_seconds())
await self.channel_layer.send(key, {
"type": value['type'],
"message": value['message']
}) | [
"async",
"def",
"emitters",
"(",
"self",
",",
"key",
",",
"value",
")",
":",
"while",
"True",
":",
"await",
"asyncio",
".",
"sleep",
"(",
"value",
"[",
"'schedule'",
"]",
".",
"total_seconds",
"(",
")",
")",
"await",
"self",
".",
"channel_layer",
".",
"send",
"(",
"key",
",",
"{",
"\"type\"",
":",
"value",
"[",
"'type'",
"]",
",",
"\"message\"",
":",
"value",
"[",
"'message'",
"]",
"}",
")"
] | Single-channel emitter | [
"Single",
"-",
"channel",
"emitter"
] | 8c653c46cdcf98398ca9d0bc6d3e47e5d621bb6a | https://github.com/rajasimon/beatserver/blob/8c653c46cdcf98398ca9d0bc6d3e47e5d621bb6a/beatserver/server.py#L40-L50 | train |
rajasimon/beatserver | beatserver/server.py | BeatServer.listener | async def listener(self, channel):
"""
Single-channel listener
"""
while True:
message = await self.channel_layer.receive(channel)
if not message.get("type", None):
raise ValueError("Worker received message with no type.")
# Make a scope and get an application instance for it
scope = {"type": "channel", "channel": channel}
instance_queue = self.get_or_create_application_instance(channel, scope)
# Run the message into the app
await instance_queue.put(message) | python | async def listener(self, channel):
"""
Single-channel listener
"""
while True:
message = await self.channel_layer.receive(channel)
if not message.get("type", None):
raise ValueError("Worker received message with no type.")
# Make a scope and get an application instance for it
scope = {"type": "channel", "channel": channel}
instance_queue = self.get_or_create_application_instance(channel, scope)
# Run the message into the app
await instance_queue.put(message) | [
"async",
"def",
"listener",
"(",
"self",
",",
"channel",
")",
":",
"while",
"True",
":",
"message",
"=",
"await",
"self",
".",
"channel_layer",
".",
"receive",
"(",
"channel",
")",
"if",
"not",
"message",
".",
"get",
"(",
"\"type\"",
",",
"None",
")",
":",
"raise",
"ValueError",
"(",
"\"Worker received message with no type.\"",
")",
"# Make a scope and get an application instance for it",
"scope",
"=",
"{",
"\"type\"",
":",
"\"channel\"",
",",
"\"channel\"",
":",
"channel",
"}",
"instance_queue",
"=",
"self",
".",
"get_or_create_application_instance",
"(",
"channel",
",",
"scope",
")",
"# Run the message into the app",
"await",
"instance_queue",
".",
"put",
"(",
"message",
")"
] | Single-channel listener | [
"Single",
"-",
"channel",
"listener"
] | 8c653c46cdcf98398ca9d0bc6d3e47e5d621bb6a | https://github.com/rajasimon/beatserver/blob/8c653c46cdcf98398ca9d0bc6d3e47e5d621bb6a/beatserver/server.py#L54-L66 | train |
pinax/pinax-ratings | pinax/ratings/templatetags/pinax_ratings_tags.py | rating_count | def rating_count(obj):
"""
Total amount of users who have submitted a positive rating for this object.
Usage:
{% rating_count obj %}
"""
count = Rating.objects.filter(
object_id=obj.pk,
content_type=ContentType.objects.get_for_model(obj),
).exclude(rating=0).count()
return count | python | def rating_count(obj):
"""
Total amount of users who have submitted a positive rating for this object.
Usage:
{% rating_count obj %}
"""
count = Rating.objects.filter(
object_id=obj.pk,
content_type=ContentType.objects.get_for_model(obj),
).exclude(rating=0).count()
return count | [
"def",
"rating_count",
"(",
"obj",
")",
":",
"count",
"=",
"Rating",
".",
"objects",
".",
"filter",
"(",
"object_id",
"=",
"obj",
".",
"pk",
",",
"content_type",
"=",
"ContentType",
".",
"objects",
".",
"get_for_model",
"(",
"obj",
")",
",",
")",
".",
"exclude",
"(",
"rating",
"=",
"0",
")",
".",
"count",
"(",
")",
"return",
"count"
] | Total amount of users who have submitted a positive rating for this object.
Usage:
{% rating_count obj %} | [
"Total",
"amount",
"of",
"users",
"who",
"have",
"submitted",
"a",
"positive",
"rating",
"for",
"this",
"object",
"."
] | eca388fea1ccd09ba844ac29a7489e41b64267f5 | https://github.com/pinax/pinax-ratings/blob/eca388fea1ccd09ba844ac29a7489e41b64267f5/pinax/ratings/templatetags/pinax_ratings_tags.py#L115-L126 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/BicolorMatrix8x8.py | BicolorMatrix8x8.set_pixel | def set_pixel(self, x, y, value):
"""Set pixel at position x, y to the given value. X and Y should be values
of 0 to 8. Value should be OFF, GREEN, RED, or YELLOW.
"""
if x < 0 or x > 7 or y < 0 or y > 7:
# Ignore out of bounds pixels.
return
# Set green LED based on 1st bit in value.
self.set_led(y * 16 + x, 1 if value & GREEN > 0 else 0)
# Set red LED based on 2nd bit in value.
self.set_led(y * 16 + x + 8, 1 if value & RED > 0 else 0) | python | def set_pixel(self, x, y, value):
"""Set pixel at position x, y to the given value. X and Y should be values
of 0 to 8. Value should be OFF, GREEN, RED, or YELLOW.
"""
if x < 0 or x > 7 or y < 0 or y > 7:
# Ignore out of bounds pixels.
return
# Set green LED based on 1st bit in value.
self.set_led(y * 16 + x, 1 if value & GREEN > 0 else 0)
# Set red LED based on 2nd bit in value.
self.set_led(y * 16 + x + 8, 1 if value & RED > 0 else 0) | [
"def",
"set_pixel",
"(",
"self",
",",
"x",
",",
"y",
",",
"value",
")",
":",
"if",
"x",
"<",
"0",
"or",
"x",
">",
"7",
"or",
"y",
"<",
"0",
"or",
"y",
">",
"7",
":",
"# Ignore out of bounds pixels.",
"return",
"# Set green LED based on 1st bit in value.",
"self",
".",
"set_led",
"(",
"y",
"*",
"16",
"+",
"x",
",",
"1",
"if",
"value",
"&",
"GREEN",
">",
"0",
"else",
"0",
")",
"# Set red LED based on 2nd bit in value.",
"self",
".",
"set_led",
"(",
"y",
"*",
"16",
"+",
"x",
"+",
"8",
",",
"1",
"if",
"value",
"&",
"RED",
">",
"0",
"else",
"0",
")"
] | Set pixel at position x, y to the given value. X and Y should be values
of 0 to 8. Value should be OFF, GREEN, RED, or YELLOW. | [
"Set",
"pixel",
"at",
"position",
"x",
"y",
"to",
"the",
"given",
"value",
".",
"X",
"and",
"Y",
"should",
"be",
"values",
"of",
"0",
"to",
"8",
".",
"Value",
"should",
"be",
"OFF",
"GREEN",
"RED",
"or",
"YELLOW",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/BicolorMatrix8x8.py#L41-L51 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/BicolorBargraph24.py | BicolorBargraph24.set_bar | def set_bar(self, bar, value):
"""Set bar to desired color. Bar should be a value of 0 to 23, and value
should be OFF, GREEN, RED, or YELLOW.
"""
if bar < 0 or bar > 23:
# Ignore out of bounds bars.
return
# Compute cathode and anode value.
c = (bar if bar < 12 else bar - 12) // 4
a = bar % 4
if bar >= 12:
a += 4
# Set green LED based on 1st bit in value.
self.set_led(c*16+a+8, 1 if value & GREEN > 0 else 0)
# Set red LED based on 2nd bit in value.
self.set_led(c*16+a, 1 if value & RED > 0 else 0) | python | def set_bar(self, bar, value):
"""Set bar to desired color. Bar should be a value of 0 to 23, and value
should be OFF, GREEN, RED, or YELLOW.
"""
if bar < 0 or bar > 23:
# Ignore out of bounds bars.
return
# Compute cathode and anode value.
c = (bar if bar < 12 else bar - 12) // 4
a = bar % 4
if bar >= 12:
a += 4
# Set green LED based on 1st bit in value.
self.set_led(c*16+a+8, 1 if value & GREEN > 0 else 0)
# Set red LED based on 2nd bit in value.
self.set_led(c*16+a, 1 if value & RED > 0 else 0) | [
"def",
"set_bar",
"(",
"self",
",",
"bar",
",",
"value",
")",
":",
"if",
"bar",
"<",
"0",
"or",
"bar",
">",
"23",
":",
"# Ignore out of bounds bars.",
"return",
"# Compute cathode and anode value.",
"c",
"=",
"(",
"bar",
"if",
"bar",
"<",
"12",
"else",
"bar",
"-",
"12",
")",
"//",
"4",
"a",
"=",
"bar",
"%",
"4",
"if",
"bar",
">=",
"12",
":",
"a",
"+=",
"4",
"# Set green LED based on 1st bit in value.",
"self",
".",
"set_led",
"(",
"c",
"*",
"16",
"+",
"a",
"+",
"8",
",",
"1",
"if",
"value",
"&",
"GREEN",
">",
"0",
"else",
"0",
")",
"# Set red LED based on 2nd bit in value.",
"self",
".",
"set_led",
"(",
"c",
"*",
"16",
"+",
"a",
",",
"1",
"if",
"value",
"&",
"RED",
">",
"0",
"else",
"0",
")"
] | Set bar to desired color. Bar should be a value of 0 to 23, and value
should be OFF, GREEN, RED, or YELLOW. | [
"Set",
"bar",
"to",
"desired",
"color",
".",
"Bar",
"should",
"be",
"a",
"value",
"of",
"0",
"to",
"23",
"and",
"value",
"should",
"be",
"OFF",
"GREEN",
"RED",
"or",
"YELLOW",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/BicolorBargraph24.py#L44-L59 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/Matrix8x8.py | Matrix8x8.animate | def animate(self, images, delay=.25):
"""Displays each of the input images in order, pausing for "delay"
seconds after each image.
Keyword arguments:
image -- An iterable collection of Image objects.
delay -- How many seconds to wait after displaying an image before
displaying the next one. (Default = .25)
"""
for image in images:
# Draw the image on the display buffer.
self.set_image(image)
# Draw the buffer to the display hardware.
self.write_display()
time.sleep(delay) | python | def animate(self, images, delay=.25):
"""Displays each of the input images in order, pausing for "delay"
seconds after each image.
Keyword arguments:
image -- An iterable collection of Image objects.
delay -- How many seconds to wait after displaying an image before
displaying the next one. (Default = .25)
"""
for image in images:
# Draw the image on the display buffer.
self.set_image(image)
# Draw the buffer to the display hardware.
self.write_display()
time.sleep(delay) | [
"def",
"animate",
"(",
"self",
",",
"images",
",",
"delay",
"=",
".25",
")",
":",
"for",
"image",
"in",
"images",
":",
"# Draw the image on the display buffer.",
"self",
".",
"set_image",
"(",
"image",
")",
"# Draw the buffer to the display hardware.",
"self",
".",
"write_display",
"(",
")",
"time",
".",
"sleep",
"(",
"delay",
")"
] | Displays each of the input images in order, pausing for "delay"
seconds after each image.
Keyword arguments:
image -- An iterable collection of Image objects.
delay -- How many seconds to wait after displaying an image before
displaying the next one. (Default = .25) | [
"Displays",
"each",
"of",
"the",
"input",
"images",
"in",
"order",
"pausing",
"for",
"delay",
"seconds",
"after",
"each",
"image",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/Matrix8x8.py#L160-L175 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/Matrix8x16.py | Matrix8x16.set_pixel | def set_pixel(self, x, y, value):
"""Set pixel at position x, y to the given value. X and Y should be values
of 0 to 7 and 0 to 15, resp. Value should be 0 for off and non-zero for on.
"""
if x < 0 or x > 7 or y < 0 or y > 15:
# Ignore out of bounds pixels.
return
self.set_led((7 - x) * 16 + y, value) | python | def set_pixel(self, x, y, value):
"""Set pixel at position x, y to the given value. X and Y should be values
of 0 to 7 and 0 to 15, resp. Value should be 0 for off and non-zero for on.
"""
if x < 0 or x > 7 or y < 0 or y > 15:
# Ignore out of bounds pixels.
return
self.set_led((7 - x) * 16 + y, value) | [
"def",
"set_pixel",
"(",
"self",
",",
"x",
",",
"y",
",",
"value",
")",
":",
"if",
"x",
"<",
"0",
"or",
"x",
">",
"7",
"or",
"y",
"<",
"0",
"or",
"y",
">",
"15",
":",
"# Ignore out of bounds pixels.",
"return",
"self",
".",
"set_led",
"(",
"(",
"7",
"-",
"x",
")",
"*",
"16",
"+",
"y",
",",
"value",
")"
] | Set pixel at position x, y to the given value. X and Y should be values
of 0 to 7 and 0 to 15, resp. Value should be 0 for off and non-zero for on. | [
"Set",
"pixel",
"at",
"position",
"x",
"y",
"to",
"the",
"given",
"value",
".",
"X",
"and",
"Y",
"should",
"be",
"values",
"of",
"0",
"to",
"7",
"and",
"0",
"to",
"15",
"resp",
".",
"Value",
"should",
"be",
"0",
"for",
"off",
"and",
"non",
"-",
"zero",
"for",
"on",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/Matrix8x16.py#L35-L42 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/Matrix8x16.py | Matrix8x16.set_image | def set_image(self, image):
"""Set display buffer to Python Image Library image. Image will be converted
to 1 bit color and non-zero color values will light the LEDs.
"""
imwidth, imheight = image.size
if imwidth != 8 or imheight != 16:
raise ValueError('Image must be an 8x16 pixels in size.')
# Convert image to 1 bit color and grab all the pixels.
pix = image.convert('1').load()
# Loop through each pixel and write the display buffer pixel.
for x in xrange(8):
for y in xrange(16):
color = pix[(x, y)]
# Handle the color of the pixel, off or on.
if color == 0:
self.set_pixel(x, y, 0)
else:
self.set_pixel(x, y, 1) | python | def set_image(self, image):
"""Set display buffer to Python Image Library image. Image will be converted
to 1 bit color and non-zero color values will light the LEDs.
"""
imwidth, imheight = image.size
if imwidth != 8 or imheight != 16:
raise ValueError('Image must be an 8x16 pixels in size.')
# Convert image to 1 bit color and grab all the pixels.
pix = image.convert('1').load()
# Loop through each pixel and write the display buffer pixel.
for x in xrange(8):
for y in xrange(16):
color = pix[(x, y)]
# Handle the color of the pixel, off or on.
if color == 0:
self.set_pixel(x, y, 0)
else:
self.set_pixel(x, y, 1) | [
"def",
"set_image",
"(",
"self",
",",
"image",
")",
":",
"imwidth",
",",
"imheight",
"=",
"image",
".",
"size",
"if",
"imwidth",
"!=",
"8",
"or",
"imheight",
"!=",
"16",
":",
"raise",
"ValueError",
"(",
"'Image must be an 8x16 pixels in size.'",
")",
"# Convert image to 1 bit color and grab all the pixels.",
"pix",
"=",
"image",
".",
"convert",
"(",
"'1'",
")",
".",
"load",
"(",
")",
"# Loop through each pixel and write the display buffer pixel.",
"for",
"x",
"in",
"xrange",
"(",
"8",
")",
":",
"for",
"y",
"in",
"xrange",
"(",
"16",
")",
":",
"color",
"=",
"pix",
"[",
"(",
"x",
",",
"y",
")",
"]",
"# Handle the color of the pixel, off or on.",
"if",
"color",
"==",
"0",
":",
"self",
".",
"set_pixel",
"(",
"x",
",",
"y",
",",
"0",
")",
"else",
":",
"self",
".",
"set_pixel",
"(",
"x",
",",
"y",
",",
"1",
")"
] | Set display buffer to Python Image Library image. Image will be converted
to 1 bit color and non-zero color values will light the LEDs. | [
"Set",
"display",
"buffer",
"to",
"Python",
"Image",
"Library",
"image",
".",
"Image",
"will",
"be",
"converted",
"to",
"1",
"bit",
"color",
"and",
"non",
"-",
"zero",
"color",
"values",
"will",
"light",
"the",
"LEDs",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/Matrix8x16.py#L44-L61 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/Matrix8x16.py | Matrix8x16.horizontal_scroll | def horizontal_scroll(self, image, padding=True):
"""Returns a list of images which appear to scroll from left to right
across the input image when displayed on the LED matrix in order.
The input image is not limited to being 8x16. If the input image is
larger than this, then all columns will be scrolled through but only
the top 16 rows of pixels will be displayed.
Keyword arguments:
image -- The image to scroll across.
padding -- If True, the animation will begin with a blank screen and the
input image will scroll into the blank screen one pixel column at a
time. Similarly, after scrolling across the whole input image, the
end of the image will scroll out of a blank screen one column at a
time. If this is not True, then only the input image will be scroll
across without beginning or ending with "whitespace."
(Default = True)
"""
image_list = list()
width = image.size[0]
# Scroll into the blank image.
if padding:
for x in range(8):
section = image.crop((0, 0, x, 16))
display_section = self.create_blank_image()
display_section.paste(section, (8 - x, 0, 8, 16))
image_list.append(display_section)
#Scroll across the input image.
for x in range(8, width + 1):
section = image.crop((x - 8, 0, x, 16))
display_section = self.create_blank_image()
display_section.paste(section, (0, 0, 8, 16))
image_list.append(display_section)
#Scroll out, leaving the blank image.
if padding:
for x in range(width - 7, width + 1):
section = image.crop((x, 0, width, 16))
display_section = self.create_blank_image()
display_section.paste(section, (0, 0, 7 - (x - (width - 7)), 16))
image_list.append(display_section)
#Return the list of images created
return image_list | python | def horizontal_scroll(self, image, padding=True):
"""Returns a list of images which appear to scroll from left to right
across the input image when displayed on the LED matrix in order.
The input image is not limited to being 8x16. If the input image is
larger than this, then all columns will be scrolled through but only
the top 16 rows of pixels will be displayed.
Keyword arguments:
image -- The image to scroll across.
padding -- If True, the animation will begin with a blank screen and the
input image will scroll into the blank screen one pixel column at a
time. Similarly, after scrolling across the whole input image, the
end of the image will scroll out of a blank screen one column at a
time. If this is not True, then only the input image will be scroll
across without beginning or ending with "whitespace."
(Default = True)
"""
image_list = list()
width = image.size[0]
# Scroll into the blank image.
if padding:
for x in range(8):
section = image.crop((0, 0, x, 16))
display_section = self.create_blank_image()
display_section.paste(section, (8 - x, 0, 8, 16))
image_list.append(display_section)
#Scroll across the input image.
for x in range(8, width + 1):
section = image.crop((x - 8, 0, x, 16))
display_section = self.create_blank_image()
display_section.paste(section, (0, 0, 8, 16))
image_list.append(display_section)
#Scroll out, leaving the blank image.
if padding:
for x in range(width - 7, width + 1):
section = image.crop((x, 0, width, 16))
display_section = self.create_blank_image()
display_section.paste(section, (0, 0, 7 - (x - (width - 7)), 16))
image_list.append(display_section)
#Return the list of images created
return image_list | [
"def",
"horizontal_scroll",
"(",
"self",
",",
"image",
",",
"padding",
"=",
"True",
")",
":",
"image_list",
"=",
"list",
"(",
")",
"width",
"=",
"image",
".",
"size",
"[",
"0",
"]",
"# Scroll into the blank image.",
"if",
"padding",
":",
"for",
"x",
"in",
"range",
"(",
"8",
")",
":",
"section",
"=",
"image",
".",
"crop",
"(",
"(",
"0",
",",
"0",
",",
"x",
",",
"16",
")",
")",
"display_section",
"=",
"self",
".",
"create_blank_image",
"(",
")",
"display_section",
".",
"paste",
"(",
"section",
",",
"(",
"8",
"-",
"x",
",",
"0",
",",
"8",
",",
"16",
")",
")",
"image_list",
".",
"append",
"(",
"display_section",
")",
"#Scroll across the input image.",
"for",
"x",
"in",
"range",
"(",
"8",
",",
"width",
"+",
"1",
")",
":",
"section",
"=",
"image",
".",
"crop",
"(",
"(",
"x",
"-",
"8",
",",
"0",
",",
"x",
",",
"16",
")",
")",
"display_section",
"=",
"self",
".",
"create_blank_image",
"(",
")",
"display_section",
".",
"paste",
"(",
"section",
",",
"(",
"0",
",",
"0",
",",
"8",
",",
"16",
")",
")",
"image_list",
".",
"append",
"(",
"display_section",
")",
"#Scroll out, leaving the blank image.",
"if",
"padding",
":",
"for",
"x",
"in",
"range",
"(",
"width",
"-",
"7",
",",
"width",
"+",
"1",
")",
":",
"section",
"=",
"image",
".",
"crop",
"(",
"(",
"x",
",",
"0",
",",
"width",
",",
"16",
")",
")",
"display_section",
"=",
"self",
".",
"create_blank_image",
"(",
")",
"display_section",
".",
"paste",
"(",
"section",
",",
"(",
"0",
",",
"0",
",",
"7",
"-",
"(",
"x",
"-",
"(",
"width",
"-",
"7",
")",
")",
",",
"16",
")",
")",
"image_list",
".",
"append",
"(",
"display_section",
")",
"#Return the list of images created",
"return",
"image_list"
] | Returns a list of images which appear to scroll from left to right
across the input image when displayed on the LED matrix in order.
The input image is not limited to being 8x16. If the input image is
larger than this, then all columns will be scrolled through but only
the top 16 rows of pixels will be displayed.
Keyword arguments:
image -- The image to scroll across.
padding -- If True, the animation will begin with a blank screen and the
input image will scroll into the blank screen one pixel column at a
time. Similarly, after scrolling across the whole input image, the
end of the image will scroll out of a blank screen one column at a
time. If this is not True, then only the input image will be scroll
across without beginning or ending with "whitespace."
(Default = True) | [
"Returns",
"a",
"list",
"of",
"images",
"which",
"appear",
"to",
"scroll",
"from",
"left",
"to",
"right",
"across",
"the",
"input",
"image",
"when",
"displayed",
"on",
"the",
"LED",
"matrix",
"in",
"order",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/Matrix8x16.py#L67-L112 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/Matrix8x16.py | Matrix8x16.vertical_scroll | def vertical_scroll(self, image, padding=True):
"""Returns a list of images which appear to scroll from top to bottom
down the input image when displayed on the LED matrix in order.
The input image is not limited to being 8x16. If the input image is
largerthan this, then all rows will be scrolled through but only the
left-most 8 columns of pixels will be displayed.
Keyword arguments:
image -- The image to scroll down.
padding -- If True, the animation will begin with a blank screen and the
input image will scroll into the blank screen one pixel row at a
time. Similarly, after scrolling down the whole input image, the end
of the image will scroll out of a blank screen one row at a time.
If this is not True, then only the input image will be scroll down
without beginning or ending with "whitespace." (Default = True)
"""
image_list = list()
height = image.size[1]
# Scroll into the blank image.
if padding:
for y in range(16):
section = image.crop((0, 0, 8, y))
display_section = self.create_blank_image()
display_section.paste(section, (0, 8 - y, 8, 16))
image_list.append(display_section)
#Scroll across the input image.
for y in range(16, height + 1):
section = image.crop((0, y - 16, 8, y))
display_section = self.create_blank_image()
display_section.paste(section, (0, 0, 8, 16))
image_list.append(display_section)
#Scroll out, leaving the blank image.
if padding:
for y in range(height - 15, height + 1):
section = image.crop((0, y, 8, height))
display_section = self.create_blank_image()
display_section.paste(section, (0, 0, 8, 7 - (y - (height - 15))))
image_list.append(display_section)
#Return the list of images created
return image_list | python | def vertical_scroll(self, image, padding=True):
"""Returns a list of images which appear to scroll from top to bottom
down the input image when displayed on the LED matrix in order.
The input image is not limited to being 8x16. If the input image is
largerthan this, then all rows will be scrolled through but only the
left-most 8 columns of pixels will be displayed.
Keyword arguments:
image -- The image to scroll down.
padding -- If True, the animation will begin with a blank screen and the
input image will scroll into the blank screen one pixel row at a
time. Similarly, after scrolling down the whole input image, the end
of the image will scroll out of a blank screen one row at a time.
If this is not True, then only the input image will be scroll down
without beginning or ending with "whitespace." (Default = True)
"""
image_list = list()
height = image.size[1]
# Scroll into the blank image.
if padding:
for y in range(16):
section = image.crop((0, 0, 8, y))
display_section = self.create_blank_image()
display_section.paste(section, (0, 8 - y, 8, 16))
image_list.append(display_section)
#Scroll across the input image.
for y in range(16, height + 1):
section = image.crop((0, y - 16, 8, y))
display_section = self.create_blank_image()
display_section.paste(section, (0, 0, 8, 16))
image_list.append(display_section)
#Scroll out, leaving the blank image.
if padding:
for y in range(height - 15, height + 1):
section = image.crop((0, y, 8, height))
display_section = self.create_blank_image()
display_section.paste(section, (0, 0, 8, 7 - (y - (height - 15))))
image_list.append(display_section)
#Return the list of images created
return image_list | [
"def",
"vertical_scroll",
"(",
"self",
",",
"image",
",",
"padding",
"=",
"True",
")",
":",
"image_list",
"=",
"list",
"(",
")",
"height",
"=",
"image",
".",
"size",
"[",
"1",
"]",
"# Scroll into the blank image.",
"if",
"padding",
":",
"for",
"y",
"in",
"range",
"(",
"16",
")",
":",
"section",
"=",
"image",
".",
"crop",
"(",
"(",
"0",
",",
"0",
",",
"8",
",",
"y",
")",
")",
"display_section",
"=",
"self",
".",
"create_blank_image",
"(",
")",
"display_section",
".",
"paste",
"(",
"section",
",",
"(",
"0",
",",
"8",
"-",
"y",
",",
"8",
",",
"16",
")",
")",
"image_list",
".",
"append",
"(",
"display_section",
")",
"#Scroll across the input image.",
"for",
"y",
"in",
"range",
"(",
"16",
",",
"height",
"+",
"1",
")",
":",
"section",
"=",
"image",
".",
"crop",
"(",
"(",
"0",
",",
"y",
"-",
"16",
",",
"8",
",",
"y",
")",
")",
"display_section",
"=",
"self",
".",
"create_blank_image",
"(",
")",
"display_section",
".",
"paste",
"(",
"section",
",",
"(",
"0",
",",
"0",
",",
"8",
",",
"16",
")",
")",
"image_list",
".",
"append",
"(",
"display_section",
")",
"#Scroll out, leaving the blank image.",
"if",
"padding",
":",
"for",
"y",
"in",
"range",
"(",
"height",
"-",
"15",
",",
"height",
"+",
"1",
")",
":",
"section",
"=",
"image",
".",
"crop",
"(",
"(",
"0",
",",
"y",
",",
"8",
",",
"height",
")",
")",
"display_section",
"=",
"self",
".",
"create_blank_image",
"(",
")",
"display_section",
".",
"paste",
"(",
"section",
",",
"(",
"0",
",",
"0",
",",
"8",
",",
"7",
"-",
"(",
"y",
"-",
"(",
"height",
"-",
"15",
")",
")",
")",
")",
"image_list",
".",
"append",
"(",
"display_section",
")",
"#Return the list of images created",
"return",
"image_list"
] | Returns a list of images which appear to scroll from top to bottom
down the input image when displayed on the LED matrix in order.
The input image is not limited to being 8x16. If the input image is
largerthan this, then all rows will be scrolled through but only the
left-most 8 columns of pixels will be displayed.
Keyword arguments:
image -- The image to scroll down.
padding -- If True, the animation will begin with a blank screen and the
input image will scroll into the blank screen one pixel row at a
time. Similarly, after scrolling down the whole input image, the end
of the image will scroll out of a blank screen one row at a time.
If this is not True, then only the input image will be scroll down
without beginning or ending with "whitespace." (Default = True) | [
"Returns",
"a",
"list",
"of",
"images",
"which",
"appear",
"to",
"scroll",
"from",
"top",
"to",
"bottom",
"down",
"the",
"input",
"image",
"when",
"displayed",
"on",
"the",
"LED",
"matrix",
"in",
"order",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/Matrix8x16.py#L114-L158 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/AlphaNum4.py | AlphaNum4.print_number_str | def print_number_str(self, value, justify_right=True):
"""Print a 4 character long string of numeric values to the display. This
function is similar to print_str but will interpret periods not as
characters but as decimal points associated with the previous character.
"""
# Calculate length of value without decimals.
length = len(value.translate(None, '.'))
# Error if value without decimals is longer than 4 characters.
if length > 4:
self.print_str('----')
return
# Calculcate starting position of digits based on justification.
pos = (4-length) if justify_right else 0
# Go through each character and print it on the display.
for i, ch in enumerate(value):
if ch == '.':
# Print decimal points on the previous digit.
self.set_decimal(pos-1, True)
else:
self.set_digit(pos, ch)
pos += 1 | python | def print_number_str(self, value, justify_right=True):
"""Print a 4 character long string of numeric values to the display. This
function is similar to print_str but will interpret periods not as
characters but as decimal points associated with the previous character.
"""
# Calculate length of value without decimals.
length = len(value.translate(None, '.'))
# Error if value without decimals is longer than 4 characters.
if length > 4:
self.print_str('----')
return
# Calculcate starting position of digits based on justification.
pos = (4-length) if justify_right else 0
# Go through each character and print it on the display.
for i, ch in enumerate(value):
if ch == '.':
# Print decimal points on the previous digit.
self.set_decimal(pos-1, True)
else:
self.set_digit(pos, ch)
pos += 1 | [
"def",
"print_number_str",
"(",
"self",
",",
"value",
",",
"justify_right",
"=",
"True",
")",
":",
"# Calculate length of value without decimals.",
"length",
"=",
"len",
"(",
"value",
".",
"translate",
"(",
"None",
",",
"'.'",
")",
")",
"# Error if value without decimals is longer than 4 characters.",
"if",
"length",
">",
"4",
":",
"self",
".",
"print_str",
"(",
"'----'",
")",
"return",
"# Calculcate starting position of digits based on justification.",
"pos",
"=",
"(",
"4",
"-",
"length",
")",
"if",
"justify_right",
"else",
"0",
"# Go through each character and print it on the display.",
"for",
"i",
",",
"ch",
"in",
"enumerate",
"(",
"value",
")",
":",
"if",
"ch",
"==",
"'.'",
":",
"# Print decimal points on the previous digit.",
"self",
".",
"set_decimal",
"(",
"pos",
"-",
"1",
",",
"True",
")",
"else",
":",
"self",
".",
"set_digit",
"(",
"pos",
",",
"ch",
")",
"pos",
"+=",
"1"
] | Print a 4 character long string of numeric values to the display. This
function is similar to print_str but will interpret periods not as
characters but as decimal points associated with the previous character. | [
"Print",
"a",
"4",
"character",
"long",
"string",
"of",
"numeric",
"values",
"to",
"the",
"display",
".",
"This",
"function",
"is",
"similar",
"to",
"print_str",
"but",
"will",
"interpret",
"periods",
"not",
"as",
"characters",
"but",
"as",
"decimal",
"points",
"associated",
"with",
"the",
"previous",
"character",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/AlphaNum4.py#L177-L197 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/AlphaNum4.py | AlphaNum4.print_float | def print_float(self, value, decimal_digits=2, justify_right=True):
"""Print a numeric value to the display. If value is negative
it will be printed with a leading minus sign. Decimal digits is the
desired number of digits after the decimal point.
"""
format_string = '{{0:0.{0}F}}'.format(decimal_digits)
self.print_number_str(format_string.format(value), justify_right) | python | def print_float(self, value, decimal_digits=2, justify_right=True):
"""Print a numeric value to the display. If value is negative
it will be printed with a leading minus sign. Decimal digits is the
desired number of digits after the decimal point.
"""
format_string = '{{0:0.{0}F}}'.format(decimal_digits)
self.print_number_str(format_string.format(value), justify_right) | [
"def",
"print_float",
"(",
"self",
",",
"value",
",",
"decimal_digits",
"=",
"2",
",",
"justify_right",
"=",
"True",
")",
":",
"format_string",
"=",
"'{{0:0.{0}F}}'",
".",
"format",
"(",
"decimal_digits",
")",
"self",
".",
"print_number_str",
"(",
"format_string",
".",
"format",
"(",
"value",
")",
",",
"justify_right",
")"
] | Print a numeric value to the display. If value is negative
it will be printed with a leading minus sign. Decimal digits is the
desired number of digits after the decimal point. | [
"Print",
"a",
"numeric",
"value",
"to",
"the",
"display",
".",
"If",
"value",
"is",
"negative",
"it",
"will",
"be",
"printed",
"with",
"a",
"leading",
"minus",
"sign",
".",
"Decimal",
"digits",
"is",
"the",
"desired",
"number",
"of",
"digits",
"after",
"the",
"decimal",
"point",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/AlphaNum4.py#L199-L205 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/SevenSegment.py | SevenSegment.set_left_colon | def set_left_colon(self, show_colon):
"""Turn the left colon on with show color True, or off with show colon
False. Only the large 1.2" 7-segment display has a left colon.
"""
if show_colon:
self.buffer[4] |= 0x04
self.buffer[4] |= 0x08
else:
self.buffer[4] &= (~0x04) & 0xFF
self.buffer[4] &= (~0x08) & 0xFF | python | def set_left_colon(self, show_colon):
"""Turn the left colon on with show color True, or off with show colon
False. Only the large 1.2" 7-segment display has a left colon.
"""
if show_colon:
self.buffer[4] |= 0x04
self.buffer[4] |= 0x08
else:
self.buffer[4] &= (~0x04) & 0xFF
self.buffer[4] &= (~0x08) & 0xFF | [
"def",
"set_left_colon",
"(",
"self",
",",
"show_colon",
")",
":",
"if",
"show_colon",
":",
"self",
".",
"buffer",
"[",
"4",
"]",
"|=",
"0x04",
"self",
".",
"buffer",
"[",
"4",
"]",
"|=",
"0x08",
"else",
":",
"self",
".",
"buffer",
"[",
"4",
"]",
"&=",
"(",
"~",
"0x04",
")",
"&",
"0xFF",
"self",
".",
"buffer",
"[",
"4",
"]",
"&=",
"(",
"~",
"0x08",
")",
"&",
"0xFF"
] | Turn the left colon on with show color True, or off with show colon
False. Only the large 1.2" 7-segment display has a left colon. | [
"Turn",
"the",
"left",
"colon",
"on",
"with",
"show",
"color",
"True",
"or",
"off",
"with",
"show",
"colon",
"False",
".",
"Only",
"the",
"large",
"1",
".",
"2",
"7",
"-",
"segment",
"display",
"has",
"a",
"left",
"colon",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/SevenSegment.py#L145-L154 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/SevenSegment.py | SevenSegment.print_number_str | def print_number_str(self, value, justify_right=True):
"""Print a 4 character long string of numeric values to the display.
Characters in the string should be any supported character by set_digit,
or a decimal point. Decimal point characters will be associated with
the previous character.
"""
# Calculate length of value without decimals.
length = sum(map(lambda x: 1 if x != '.' else 0, value))
# Error if value without decimals is longer than 4 characters.
if length > 4:
self.print_number_str('----')
return
# Calculcate starting position of digits based on justification.
pos = (4-length) if justify_right else 0
# Go through each character and print it on the display.
for i, ch in enumerate(value):
if ch == '.':
# Print decimal points on the previous digit.
self.set_decimal(pos-1, True)
else:
self.set_digit(pos, ch)
pos += 1 | python | def print_number_str(self, value, justify_right=True):
"""Print a 4 character long string of numeric values to the display.
Characters in the string should be any supported character by set_digit,
or a decimal point. Decimal point characters will be associated with
the previous character.
"""
# Calculate length of value without decimals.
length = sum(map(lambda x: 1 if x != '.' else 0, value))
# Error if value without decimals is longer than 4 characters.
if length > 4:
self.print_number_str('----')
return
# Calculcate starting position of digits based on justification.
pos = (4-length) if justify_right else 0
# Go through each character and print it on the display.
for i, ch in enumerate(value):
if ch == '.':
# Print decimal points on the previous digit.
self.set_decimal(pos-1, True)
else:
self.set_digit(pos, ch)
pos += 1 | [
"def",
"print_number_str",
"(",
"self",
",",
"value",
",",
"justify_right",
"=",
"True",
")",
":",
"# Calculate length of value without decimals.",
"length",
"=",
"sum",
"(",
"map",
"(",
"lambda",
"x",
":",
"1",
"if",
"x",
"!=",
"'.'",
"else",
"0",
",",
"value",
")",
")",
"# Error if value without decimals is longer than 4 characters.",
"if",
"length",
">",
"4",
":",
"self",
".",
"print_number_str",
"(",
"'----'",
")",
"return",
"# Calculcate starting position of digits based on justification.",
"pos",
"=",
"(",
"4",
"-",
"length",
")",
"if",
"justify_right",
"else",
"0",
"# Go through each character and print it on the display.",
"for",
"i",
",",
"ch",
"in",
"enumerate",
"(",
"value",
")",
":",
"if",
"ch",
"==",
"'.'",
":",
"# Print decimal points on the previous digit.",
"self",
".",
"set_decimal",
"(",
"pos",
"-",
"1",
",",
"True",
")",
"else",
":",
"self",
".",
"set_digit",
"(",
"pos",
",",
"ch",
")",
"pos",
"+=",
"1"
] | Print a 4 character long string of numeric values to the display.
Characters in the string should be any supported character by set_digit,
or a decimal point. Decimal point characters will be associated with
the previous character. | [
"Print",
"a",
"4",
"character",
"long",
"string",
"of",
"numeric",
"values",
"to",
"the",
"display",
".",
"Characters",
"in",
"the",
"string",
"should",
"be",
"any",
"supported",
"character",
"by",
"set_digit",
"or",
"a",
"decimal",
"point",
".",
"Decimal",
"point",
"characters",
"will",
"be",
"associated",
"with",
"the",
"previous",
"character",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/SevenSegment.py#L167-L188 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/HT16K33.py | HT16K33.begin | def begin(self):
"""Initialize driver with LEDs enabled and all turned off."""
# Turn on the oscillator.
self._device.writeList(HT16K33_SYSTEM_SETUP | HT16K33_OSCILLATOR, [])
# Turn display on with no blinking.
self.set_blink(HT16K33_BLINK_OFF)
# Set display to full brightness.
self.set_brightness(15) | python | def begin(self):
"""Initialize driver with LEDs enabled and all turned off."""
# Turn on the oscillator.
self._device.writeList(HT16K33_SYSTEM_SETUP | HT16K33_OSCILLATOR, [])
# Turn display on with no blinking.
self.set_blink(HT16K33_BLINK_OFF)
# Set display to full brightness.
self.set_brightness(15) | [
"def",
"begin",
"(",
"self",
")",
":",
"# Turn on the oscillator.",
"self",
".",
"_device",
".",
"writeList",
"(",
"HT16K33_SYSTEM_SETUP",
"|",
"HT16K33_OSCILLATOR",
",",
"[",
"]",
")",
"# Turn display on with no blinking.",
"self",
".",
"set_blink",
"(",
"HT16K33_BLINK_OFF",
")",
"# Set display to full brightness.",
"self",
".",
"set_brightness",
"(",
"15",
")"
] | Initialize driver with LEDs enabled and all turned off. | [
"Initialize",
"driver",
"with",
"LEDs",
"enabled",
"and",
"all",
"turned",
"off",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/HT16K33.py#L50-L57 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/HT16K33.py | HT16K33.write_display | def write_display(self):
"""Write display buffer to display hardware."""
for i, value in enumerate(self.buffer):
self._device.write8(i, value) | python | def write_display(self):
"""Write display buffer to display hardware."""
for i, value in enumerate(self.buffer):
self._device.write8(i, value) | [
"def",
"write_display",
"(",
"self",
")",
":",
"for",
"i",
",",
"value",
"in",
"enumerate",
"(",
"self",
".",
"buffer",
")",
":",
"self",
".",
"_device",
".",
"write8",
"(",
"i",
",",
"value",
")"
] | Write display buffer to display hardware. | [
"Write",
"display",
"buffer",
"to",
"display",
"hardware",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/HT16K33.py#L93-L96 | train |
adafruit/Adafruit_Python_LED_Backpack | Adafruit_LED_Backpack/HT16K33.py | HT16K33.clear | def clear(self):
"""Clear contents of display buffer."""
for i, value in enumerate(self.buffer):
self.buffer[i] = 0 | python | def clear(self):
"""Clear contents of display buffer."""
for i, value in enumerate(self.buffer):
self.buffer[i] = 0 | [
"def",
"clear",
"(",
"self",
")",
":",
"for",
"i",
",",
"value",
"in",
"enumerate",
"(",
"self",
".",
"buffer",
")",
":",
"self",
".",
"buffer",
"[",
"i",
"]",
"=",
"0"
] | Clear contents of display buffer. | [
"Clear",
"contents",
"of",
"display",
"buffer",
"."
] | 7356b4dd8b4bb162d60987878c2cb752fdd017d5 | https://github.com/adafruit/Adafruit_Python_LED_Backpack/blob/7356b4dd8b4bb162d60987878c2cb752fdd017d5/Adafruit_LED_Backpack/HT16K33.py#L98-L101 | train |
swisscom/cleanerversion | versions/admin.py | VersionedAdmin.get_readonly_fields | def get_readonly_fields(self, request, obj=None):
"""
This is required a subclass of VersionedAdmin has readonly_fields
ours won't be undone
"""
if obj:
return list(self.readonly_fields) + ['id', 'identity',
'is_current']
return self.readonly_fields | python | def get_readonly_fields(self, request, obj=None):
"""
This is required a subclass of VersionedAdmin has readonly_fields
ours won't be undone
"""
if obj:
return list(self.readonly_fields) + ['id', 'identity',
'is_current']
return self.readonly_fields | [
"def",
"get_readonly_fields",
"(",
"self",
",",
"request",
",",
"obj",
"=",
"None",
")",
":",
"if",
"obj",
":",
"return",
"list",
"(",
"self",
".",
"readonly_fields",
")",
"+",
"[",
"'id'",
",",
"'identity'",
",",
"'is_current'",
"]",
"return",
"self",
".",
"readonly_fields"
] | This is required a subclass of VersionedAdmin has readonly_fields
ours won't be undone | [
"This",
"is",
"required",
"a",
"subclass",
"of",
"VersionedAdmin",
"has",
"readonly_fields",
"ours",
"won",
"t",
"be",
"undone"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/admin.py#L144-L152 | train |
swisscom/cleanerversion | versions/admin.py | VersionedAdmin.get_list_display | def get_list_display(self, request):
"""
This method determines which fields go in the changelist
"""
# Force cast to list as super get_list_display could return a tuple
list_display = list(
super(VersionedAdmin, self).get_list_display(request))
# Preprend the following fields to list display
if self.list_display_show_identity:
list_display = ['identity_shortener', ] + list_display
# Append the following fields to list display
if self.list_display_show_start_date:
list_display += ['version_start_date', ]
if self.list_display_show_end_date:
list_display += ['version_end_date', ]
return list_display + ['is_current', ] | python | def get_list_display(self, request):
"""
This method determines which fields go in the changelist
"""
# Force cast to list as super get_list_display could return a tuple
list_display = list(
super(VersionedAdmin, self).get_list_display(request))
# Preprend the following fields to list display
if self.list_display_show_identity:
list_display = ['identity_shortener', ] + list_display
# Append the following fields to list display
if self.list_display_show_start_date:
list_display += ['version_start_date', ]
if self.list_display_show_end_date:
list_display += ['version_end_date', ]
return list_display + ['is_current', ] | [
"def",
"get_list_display",
"(",
"self",
",",
"request",
")",
":",
"# Force cast to list as super get_list_display could return a tuple",
"list_display",
"=",
"list",
"(",
"super",
"(",
"VersionedAdmin",
",",
"self",
")",
".",
"get_list_display",
"(",
"request",
")",
")",
"# Preprend the following fields to list display",
"if",
"self",
".",
"list_display_show_identity",
":",
"list_display",
"=",
"[",
"'identity_shortener'",
",",
"]",
"+",
"list_display",
"# Append the following fields to list display",
"if",
"self",
".",
"list_display_show_start_date",
":",
"list_display",
"+=",
"[",
"'version_start_date'",
",",
"]",
"if",
"self",
".",
"list_display_show_end_date",
":",
"list_display",
"+=",
"[",
"'version_end_date'",
",",
"]",
"return",
"list_display",
"+",
"[",
"'is_current'",
",",
"]"
] | This method determines which fields go in the changelist | [
"This",
"method",
"determines",
"which",
"fields",
"go",
"in",
"the",
"changelist"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/admin.py#L157-L176 | train |
swisscom/cleanerversion | versions/admin.py | VersionedAdmin.get_list_filter | def get_list_filter(self, request):
"""
Adds versionable custom filtering ability to changelist
"""
list_filter = super(VersionedAdmin, self).get_list_filter(request)
return list(list_filter) + [('version_start_date', DateTimeFilter),
IsCurrentFilter] | python | def get_list_filter(self, request):
"""
Adds versionable custom filtering ability to changelist
"""
list_filter = super(VersionedAdmin, self).get_list_filter(request)
return list(list_filter) + [('version_start_date', DateTimeFilter),
IsCurrentFilter] | [
"def",
"get_list_filter",
"(",
"self",
",",
"request",
")",
":",
"list_filter",
"=",
"super",
"(",
"VersionedAdmin",
",",
"self",
")",
".",
"get_list_filter",
"(",
"request",
")",
"return",
"list",
"(",
"list_filter",
")",
"+",
"[",
"(",
"'version_start_date'",
",",
"DateTimeFilter",
")",
",",
"IsCurrentFilter",
"]"
] | Adds versionable custom filtering ability to changelist | [
"Adds",
"versionable",
"custom",
"filtering",
"ability",
"to",
"changelist"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/admin.py#L178-L184 | train |
swisscom/cleanerversion | versions/admin.py | VersionedAdmin.restore | def restore(self, request, *args, **kwargs):
"""
View for restoring object from change view
"""
paths = request.path_info.split('/')
object_id_index = paths.index("restore") - 2
object_id = paths[object_id_index]
obj = super(VersionedAdmin, self).get_object(request, object_id)
obj.restore()
admin_wordIndex = object_id_index - 3
path = "/%s" % ("/".join(paths[admin_wordIndex:object_id_index]))
opts = self.model._meta
msg_dict = {
'name': force_text(opts.verbose_name),
'obj': format_html('<a href="{}">{}</a>',
urlquote(request.path), obj),
}
msg = format_html(_('The {name} "{obj}" was restored successfully.'),
**msg_dict)
self.message_user(request, msg, messages.SUCCESS)
return HttpResponseRedirect(path) | python | def restore(self, request, *args, **kwargs):
"""
View for restoring object from change view
"""
paths = request.path_info.split('/')
object_id_index = paths.index("restore") - 2
object_id = paths[object_id_index]
obj = super(VersionedAdmin, self).get_object(request, object_id)
obj.restore()
admin_wordIndex = object_id_index - 3
path = "/%s" % ("/".join(paths[admin_wordIndex:object_id_index]))
opts = self.model._meta
msg_dict = {
'name': force_text(opts.verbose_name),
'obj': format_html('<a href="{}">{}</a>',
urlquote(request.path), obj),
}
msg = format_html(_('The {name} "{obj}" was restored successfully.'),
**msg_dict)
self.message_user(request, msg, messages.SUCCESS)
return HttpResponseRedirect(path) | [
"def",
"restore",
"(",
"self",
",",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"paths",
"=",
"request",
".",
"path_info",
".",
"split",
"(",
"'/'",
")",
"object_id_index",
"=",
"paths",
".",
"index",
"(",
"\"restore\"",
")",
"-",
"2",
"object_id",
"=",
"paths",
"[",
"object_id_index",
"]",
"obj",
"=",
"super",
"(",
"VersionedAdmin",
",",
"self",
")",
".",
"get_object",
"(",
"request",
",",
"object_id",
")",
"obj",
".",
"restore",
"(",
")",
"admin_wordIndex",
"=",
"object_id_index",
"-",
"3",
"path",
"=",
"\"/%s\"",
"%",
"(",
"\"/\"",
".",
"join",
"(",
"paths",
"[",
"admin_wordIndex",
":",
"object_id_index",
"]",
")",
")",
"opts",
"=",
"self",
".",
"model",
".",
"_meta",
"msg_dict",
"=",
"{",
"'name'",
":",
"force_text",
"(",
"opts",
".",
"verbose_name",
")",
",",
"'obj'",
":",
"format_html",
"(",
"'<a href=\"{}\">{}</a>'",
",",
"urlquote",
"(",
"request",
".",
"path",
")",
",",
"obj",
")",
",",
"}",
"msg",
"=",
"format_html",
"(",
"_",
"(",
"'The {name} \"{obj}\" was restored successfully.'",
")",
",",
"*",
"*",
"msg_dict",
")",
"self",
".",
"message_user",
"(",
"request",
",",
"msg",
",",
"messages",
".",
"SUCCESS",
")",
"return",
"HttpResponseRedirect",
"(",
"path",
")"
] | View for restoring object from change view | [
"View",
"for",
"restoring",
"object",
"from",
"change",
"view"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/admin.py#L186-L209 | train |
swisscom/cleanerversion | versions/admin.py | VersionedAdmin.will_not_clone | def will_not_clone(self, request, *args, **kwargs):
"""
Add save but not clone capability in the changeview
"""
paths = request.path_info.split('/')
index_of_object_id = paths.index("will_not_clone") - 1
object_id = paths[index_of_object_id]
self.change_view(request, object_id)
admin_wordInUrl = index_of_object_id - 3
# This gets the adminsite for the app, and the model name and joins
# together with /
path = '/' + '/'.join(paths[admin_wordInUrl:index_of_object_id])
return HttpResponseRedirect(path) | python | def will_not_clone(self, request, *args, **kwargs):
"""
Add save but not clone capability in the changeview
"""
paths = request.path_info.split('/')
index_of_object_id = paths.index("will_not_clone") - 1
object_id = paths[index_of_object_id]
self.change_view(request, object_id)
admin_wordInUrl = index_of_object_id - 3
# This gets the adminsite for the app, and the model name and joins
# together with /
path = '/' + '/'.join(paths[admin_wordInUrl:index_of_object_id])
return HttpResponseRedirect(path) | [
"def",
"will_not_clone",
"(",
"self",
",",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"paths",
"=",
"request",
".",
"path_info",
".",
"split",
"(",
"'/'",
")",
"index_of_object_id",
"=",
"paths",
".",
"index",
"(",
"\"will_not_clone\"",
")",
"-",
"1",
"object_id",
"=",
"paths",
"[",
"index_of_object_id",
"]",
"self",
".",
"change_view",
"(",
"request",
",",
"object_id",
")",
"admin_wordInUrl",
"=",
"index_of_object_id",
"-",
"3",
"# This gets the adminsite for the app, and the model name and joins",
"# together with /",
"path",
"=",
"'/'",
"+",
"'/'",
".",
"join",
"(",
"paths",
"[",
"admin_wordInUrl",
":",
"index_of_object_id",
"]",
")",
"return",
"HttpResponseRedirect",
"(",
"path",
")"
] | Add save but not clone capability in the changeview | [
"Add",
"save",
"but",
"not",
"clone",
"capability",
"in",
"the",
"changeview"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/admin.py#L211-L224 | train |
swisscom/cleanerversion | versions/admin.py | VersionedAdmin.exclude | def exclude(self):
"""
Custom descriptor for exclude since there is no get_exclude method to
be overridden
"""
exclude = self.VERSIONED_EXCLUDE
if super(VersionedAdmin, self).exclude is not None:
# Force cast to list as super exclude could return a tuple
exclude = list(super(VersionedAdmin, self).exclude) + exclude
return exclude | python | def exclude(self):
"""
Custom descriptor for exclude since there is no get_exclude method to
be overridden
"""
exclude = self.VERSIONED_EXCLUDE
if super(VersionedAdmin, self).exclude is not None:
# Force cast to list as super exclude could return a tuple
exclude = list(super(VersionedAdmin, self).exclude) + exclude
return exclude | [
"def",
"exclude",
"(",
"self",
")",
":",
"exclude",
"=",
"self",
".",
"VERSIONED_EXCLUDE",
"if",
"super",
"(",
"VersionedAdmin",
",",
"self",
")",
".",
"exclude",
"is",
"not",
"None",
":",
"# Force cast to list as super exclude could return a tuple",
"exclude",
"=",
"list",
"(",
"super",
"(",
"VersionedAdmin",
",",
"self",
")",
".",
"exclude",
")",
"+",
"exclude",
"return",
"exclude"
] | Custom descriptor for exclude since there is no get_exclude method to
be overridden | [
"Custom",
"descriptor",
"for",
"exclude",
"since",
"there",
"is",
"no",
"get_exclude",
"method",
"to",
"be",
"overridden"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/admin.py#L227-L238 | train |
swisscom/cleanerversion | versions/admin.py | VersionedAdmin.get_object | def get_object(self, request, object_id, from_field=None):
"""
our implementation of get_object allows for cloning when updating an
object, not cloning when the button 'save but not clone' is pushed
and at no other time will clone be called
"""
# from_field breaks in 1.7.8
obj = super(VersionedAdmin, self).get_object(request,
object_id)
# Only clone if update view as get_object() is also called for change,
# delete, and history views
if request.method == 'POST' and \
obj and \
obj.is_latest and \
'will_not_clone' not in request.path and \
'delete' not in request.path and \
'restore' not in request.path:
obj = obj.clone()
return obj | python | def get_object(self, request, object_id, from_field=None):
"""
our implementation of get_object allows for cloning when updating an
object, not cloning when the button 'save but not clone' is pushed
and at no other time will clone be called
"""
# from_field breaks in 1.7.8
obj = super(VersionedAdmin, self).get_object(request,
object_id)
# Only clone if update view as get_object() is also called for change,
# delete, and history views
if request.method == 'POST' and \
obj and \
obj.is_latest and \
'will_not_clone' not in request.path and \
'delete' not in request.path and \
'restore' not in request.path:
obj = obj.clone()
return obj | [
"def",
"get_object",
"(",
"self",
",",
"request",
",",
"object_id",
",",
"from_field",
"=",
"None",
")",
":",
"# from_field breaks in 1.7.8",
"obj",
"=",
"super",
"(",
"VersionedAdmin",
",",
"self",
")",
".",
"get_object",
"(",
"request",
",",
"object_id",
")",
"# Only clone if update view as get_object() is also called for change,",
"# delete, and history views",
"if",
"request",
".",
"method",
"==",
"'POST'",
"and",
"obj",
"and",
"obj",
".",
"is_latest",
"and",
"'will_not_clone'",
"not",
"in",
"request",
".",
"path",
"and",
"'delete'",
"not",
"in",
"request",
".",
"path",
"and",
"'restore'",
"not",
"in",
"request",
".",
"path",
":",
"obj",
"=",
"obj",
".",
"clone",
"(",
")",
"return",
"obj"
] | our implementation of get_object allows for cloning when updating an
object, not cloning when the button 'save but not clone' is pushed
and at no other time will clone be called | [
"our",
"implementation",
"of",
"get_object",
"allows",
"for",
"cloning",
"when",
"updating",
"an",
"object",
"not",
"cloning",
"when",
"the",
"button",
"save",
"but",
"not",
"clone",
"is",
"pushed",
"and",
"at",
"no",
"other",
"time",
"will",
"clone",
"be",
"called"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/admin.py#L240-L259 | train |
swisscom/cleanerversion | versions/admin.py | VersionedAdmin.get_urls | def get_urls(self):
"""
Appends the custom will_not_clone url to the admin site
"""
not_clone_url = [url(r'^(.+)/will_not_clone/$',
admin.site.admin_view(self.will_not_clone))]
restore_url = [
url(r'^(.+)/restore/$', admin.site.admin_view(self.restore))]
return not_clone_url + restore_url + super(VersionedAdmin,
self).get_urls() | python | def get_urls(self):
"""
Appends the custom will_not_clone url to the admin site
"""
not_clone_url = [url(r'^(.+)/will_not_clone/$',
admin.site.admin_view(self.will_not_clone))]
restore_url = [
url(r'^(.+)/restore/$', admin.site.admin_view(self.restore))]
return not_clone_url + restore_url + super(VersionedAdmin,
self).get_urls() | [
"def",
"get_urls",
"(",
"self",
")",
":",
"not_clone_url",
"=",
"[",
"url",
"(",
"r'^(.+)/will_not_clone/$'",
",",
"admin",
".",
"site",
".",
"admin_view",
"(",
"self",
".",
"will_not_clone",
")",
")",
"]",
"restore_url",
"=",
"[",
"url",
"(",
"r'^(.+)/restore/$'",
",",
"admin",
".",
"site",
".",
"admin_view",
"(",
"self",
".",
"restore",
")",
")",
"]",
"return",
"not_clone_url",
"+",
"restore_url",
"+",
"super",
"(",
"VersionedAdmin",
",",
"self",
")",
".",
"get_urls",
"(",
")"
] | Appends the custom will_not_clone url to the admin site | [
"Appends",
"the",
"custom",
"will_not_clone",
"url",
"to",
"the",
"admin",
"site"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/admin.py#L297-L306 | train |
swisscom/cleanerversion | versions/util/postgresql.py | create_current_version_unique_identity_indexes | def create_current_version_unique_identity_indexes(app_name, database=None):
"""
Add partial unique indexes for the the identity column of versionable
models.
This enforces that no two *current* versions can have the same identity.
This will only try to create indexes if they do not exist in the database,
so it should be safe to run in a post_migrate signal handler. Running it
several times should leave the database in the same state as running it
once.
:param str app_name: application name whose Versionable models will be
acted on.
:param str database: database alias to use. If None, use default
connection.
:return: number of partial unique indexes created
:rtype: int
"""
indexes_created = 0
connection = database_connection(database)
with connection.cursor() as cursor:
for model in versionable_models(app_name):
if getattr(model._meta, 'managed', True):
table_name = model._meta.db_table
index_name = '%s_%s_identity_v_uniq' % (app_name, table_name)
if not index_exists(cursor, index_name):
cursor.execute(
"CREATE UNIQUE INDEX %s ON %s(%s) "
"WHERE version_end_date IS NULL"
% (index_name, table_name, 'identity'))
indexes_created += 1
return indexes_created | python | def create_current_version_unique_identity_indexes(app_name, database=None):
"""
Add partial unique indexes for the the identity column of versionable
models.
This enforces that no two *current* versions can have the same identity.
This will only try to create indexes if they do not exist in the database,
so it should be safe to run in a post_migrate signal handler. Running it
several times should leave the database in the same state as running it
once.
:param str app_name: application name whose Versionable models will be
acted on.
:param str database: database alias to use. If None, use default
connection.
:return: number of partial unique indexes created
:rtype: int
"""
indexes_created = 0
connection = database_connection(database)
with connection.cursor() as cursor:
for model in versionable_models(app_name):
if getattr(model._meta, 'managed', True):
table_name = model._meta.db_table
index_name = '%s_%s_identity_v_uniq' % (app_name, table_name)
if not index_exists(cursor, index_name):
cursor.execute(
"CREATE UNIQUE INDEX %s ON %s(%s) "
"WHERE version_end_date IS NULL"
% (index_name, table_name, 'identity'))
indexes_created += 1
return indexes_created | [
"def",
"create_current_version_unique_identity_indexes",
"(",
"app_name",
",",
"database",
"=",
"None",
")",
":",
"indexes_created",
"=",
"0",
"connection",
"=",
"database_connection",
"(",
"database",
")",
"with",
"connection",
".",
"cursor",
"(",
")",
"as",
"cursor",
":",
"for",
"model",
"in",
"versionable_models",
"(",
"app_name",
")",
":",
"if",
"getattr",
"(",
"model",
".",
"_meta",
",",
"'managed'",
",",
"True",
")",
":",
"table_name",
"=",
"model",
".",
"_meta",
".",
"db_table",
"index_name",
"=",
"'%s_%s_identity_v_uniq'",
"%",
"(",
"app_name",
",",
"table_name",
")",
"if",
"not",
"index_exists",
"(",
"cursor",
",",
"index_name",
")",
":",
"cursor",
".",
"execute",
"(",
"\"CREATE UNIQUE INDEX %s ON %s(%s) \"",
"\"WHERE version_end_date IS NULL\"",
"%",
"(",
"index_name",
",",
"table_name",
",",
"'identity'",
")",
")",
"indexes_created",
"+=",
"1",
"return",
"indexes_created"
] | Add partial unique indexes for the the identity column of versionable
models.
This enforces that no two *current* versions can have the same identity.
This will only try to create indexes if they do not exist in the database,
so it should be safe to run in a post_migrate signal handler. Running it
several times should leave the database in the same state as running it
once.
:param str app_name: application name whose Versionable models will be
acted on.
:param str database: database alias to use. If None, use default
connection.
:return: number of partial unique indexes created
:rtype: int | [
"Add",
"partial",
"unique",
"indexes",
"for",
"the",
"the",
"identity",
"column",
"of",
"versionable",
"models",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/util/postgresql.py#L148-L182 | train |
swisscom/cleanerversion | versions/models.py | VersionManager.get_queryset | def get_queryset(self):
"""
Returns a VersionedQuerySet capable of handling version time
restrictions.
:return: VersionedQuerySet
"""
qs = VersionedQuerySet(self.model, using=self._db)
if hasattr(self, 'instance') and hasattr(self.instance, '_querytime'):
qs.querytime = self.instance._querytime
return qs | python | def get_queryset(self):
"""
Returns a VersionedQuerySet capable of handling version time
restrictions.
:return: VersionedQuerySet
"""
qs = VersionedQuerySet(self.model, using=self._db)
if hasattr(self, 'instance') and hasattr(self.instance, '_querytime'):
qs.querytime = self.instance._querytime
return qs | [
"def",
"get_queryset",
"(",
"self",
")",
":",
"qs",
"=",
"VersionedQuerySet",
"(",
"self",
".",
"model",
",",
"using",
"=",
"self",
".",
"_db",
")",
"if",
"hasattr",
"(",
"self",
",",
"'instance'",
")",
"and",
"hasattr",
"(",
"self",
".",
"instance",
",",
"'_querytime'",
")",
":",
"qs",
".",
"querytime",
"=",
"self",
".",
"instance",
".",
"_querytime",
"return",
"qs"
] | Returns a VersionedQuerySet capable of handling version time
restrictions.
:return: VersionedQuerySet | [
"Returns",
"a",
"VersionedQuerySet",
"capable",
"of",
"handling",
"version",
"time",
"restrictions",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L62-L72 | train |
swisscom/cleanerversion | versions/models.py | VersionManager.next_version | def next_version(self, object, relations_as_of='end'):
"""
Return the next version of the given object.
In case there is no next object existing, meaning the given
object is the current version, the function returns this version.
Note that if object's version_end_date is None, this does not check
the database to see if there is a newer version (perhaps created by
some other code), it simply returns the passed object.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations. See ``VersionManager.version_as_of`` for details
on valid ``relations_as_of`` values.
:param Versionable object: object whose next version will be returned.
:param mixed relations_as_of: determines point in time used to access
relations. 'start'|'end'|datetime|None
:return: Versionable
"""
if object.version_end_date is None:
next = object
else:
next = self.filter(
Q(identity=object.identity),
Q(version_start_date__gte=object.version_end_date)
).order_by('version_start_date').first()
if not next:
raise ObjectDoesNotExist(
"next_version couldn't find a next version of object " +
str(object.identity))
return self.adjust_version_as_of(next, relations_as_of) | python | def next_version(self, object, relations_as_of='end'):
"""
Return the next version of the given object.
In case there is no next object existing, meaning the given
object is the current version, the function returns this version.
Note that if object's version_end_date is None, this does not check
the database to see if there is a newer version (perhaps created by
some other code), it simply returns the passed object.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations. See ``VersionManager.version_as_of`` for details
on valid ``relations_as_of`` values.
:param Versionable object: object whose next version will be returned.
:param mixed relations_as_of: determines point in time used to access
relations. 'start'|'end'|datetime|None
:return: Versionable
"""
if object.version_end_date is None:
next = object
else:
next = self.filter(
Q(identity=object.identity),
Q(version_start_date__gte=object.version_end_date)
).order_by('version_start_date').first()
if not next:
raise ObjectDoesNotExist(
"next_version couldn't find a next version of object " +
str(object.identity))
return self.adjust_version_as_of(next, relations_as_of) | [
"def",
"next_version",
"(",
"self",
",",
"object",
",",
"relations_as_of",
"=",
"'end'",
")",
":",
"if",
"object",
".",
"version_end_date",
"is",
"None",
":",
"next",
"=",
"object",
"else",
":",
"next",
"=",
"self",
".",
"filter",
"(",
"Q",
"(",
"identity",
"=",
"object",
".",
"identity",
")",
",",
"Q",
"(",
"version_start_date__gte",
"=",
"object",
".",
"version_end_date",
")",
")",
".",
"order_by",
"(",
"'version_start_date'",
")",
".",
"first",
"(",
")",
"if",
"not",
"next",
":",
"raise",
"ObjectDoesNotExist",
"(",
"\"next_version couldn't find a next version of object \"",
"+",
"str",
"(",
"object",
".",
"identity",
")",
")",
"return",
"self",
".",
"adjust_version_as_of",
"(",
"next",
",",
"relations_as_of",
")"
] | Return the next version of the given object.
In case there is no next object existing, meaning the given
object is the current version, the function returns this version.
Note that if object's version_end_date is None, this does not check
the database to see if there is a newer version (perhaps created by
some other code), it simply returns the passed object.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations. See ``VersionManager.version_as_of`` for details
on valid ``relations_as_of`` values.
:param Versionable object: object whose next version will be returned.
:param mixed relations_as_of: determines point in time used to access
relations. 'start'|'end'|datetime|None
:return: Versionable | [
"Return",
"the",
"next",
"version",
"of",
"the",
"given",
"object",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L83-L117 | train |
swisscom/cleanerversion | versions/models.py | VersionManager.previous_version | def previous_version(self, object, relations_as_of='end'):
"""
Return the previous version of the given object.
In case there is no previous object existing, meaning the given object
is the first version of the object, then the function returns this
version.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations. See ``VersionManager.version_as_of`` for details on
valid ``relations_as_of`` values.
:param Versionable object: object whose previous version will be
returned.
:param mixed relations_as_of: determines point in time used to access
relations. 'start'|'end'|datetime|None
:return: Versionable
"""
if object.version_birth_date == object.version_start_date:
previous = object
else:
previous = self.filter(
Q(identity=object.identity),
Q(version_end_date__lte=object.version_start_date)
).order_by('-version_end_date').first()
if not previous:
raise ObjectDoesNotExist(
"previous_version couldn't find a previous version of "
"object " + str(object.identity))
return self.adjust_version_as_of(previous, relations_as_of) | python | def previous_version(self, object, relations_as_of='end'):
"""
Return the previous version of the given object.
In case there is no previous object existing, meaning the given object
is the first version of the object, then the function returns this
version.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations. See ``VersionManager.version_as_of`` for details on
valid ``relations_as_of`` values.
:param Versionable object: object whose previous version will be
returned.
:param mixed relations_as_of: determines point in time used to access
relations. 'start'|'end'|datetime|None
:return: Versionable
"""
if object.version_birth_date == object.version_start_date:
previous = object
else:
previous = self.filter(
Q(identity=object.identity),
Q(version_end_date__lte=object.version_start_date)
).order_by('-version_end_date').first()
if not previous:
raise ObjectDoesNotExist(
"previous_version couldn't find a previous version of "
"object " + str(object.identity))
return self.adjust_version_as_of(previous, relations_as_of) | [
"def",
"previous_version",
"(",
"self",
",",
"object",
",",
"relations_as_of",
"=",
"'end'",
")",
":",
"if",
"object",
".",
"version_birth_date",
"==",
"object",
".",
"version_start_date",
":",
"previous",
"=",
"object",
"else",
":",
"previous",
"=",
"self",
".",
"filter",
"(",
"Q",
"(",
"identity",
"=",
"object",
".",
"identity",
")",
",",
"Q",
"(",
"version_end_date__lte",
"=",
"object",
".",
"version_start_date",
")",
")",
".",
"order_by",
"(",
"'-version_end_date'",
")",
".",
"first",
"(",
")",
"if",
"not",
"previous",
":",
"raise",
"ObjectDoesNotExist",
"(",
"\"previous_version couldn't find a previous version of \"",
"\"object \"",
"+",
"str",
"(",
"object",
".",
"identity",
")",
")",
"return",
"self",
".",
"adjust_version_as_of",
"(",
"previous",
",",
"relations_as_of",
")"
] | Return the previous version of the given object.
In case there is no previous object existing, meaning the given object
is the first version of the object, then the function returns this
version.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations. See ``VersionManager.version_as_of`` for details on
valid ``relations_as_of`` values.
:param Versionable object: object whose previous version will be
returned.
:param mixed relations_as_of: determines point in time used to access
relations. 'start'|'end'|datetime|None
:return: Versionable | [
"Return",
"the",
"previous",
"version",
"of",
"the",
"given",
"object",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L119-L151 | train |
swisscom/cleanerversion | versions/models.py | VersionManager.current_version | def current_version(self, object, relations_as_of=None, check_db=False):
"""
Return the current version of the given object.
The current version is the one having its version_end_date set to NULL.
If there is not such a version then it means the object has been
'deleted' and so there is no current version available. In this case
the function returns None.
Note that if check_db is False and object's version_end_date is None,
this does not check the database to see if there is a newer version
(perhaps created by some other code), it simply returns the passed
object.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations. See ``VersionManager.version_as_of`` for details on
valid ``relations_as_of`` values.
:param Versionable object: object whose current version will be
returned.
:param mixed relations_as_of: determines point in time used to access
relations. 'start'|'end'|datetime|None
:param bool check_db: Whether or not to look in the database for a
more recent version
:return: Versionable
"""
if object.version_end_date is None and not check_db:
current = object
else:
current = self.current.filter(identity=object.identity).first()
return self.adjust_version_as_of(current, relations_as_of) | python | def current_version(self, object, relations_as_of=None, check_db=False):
"""
Return the current version of the given object.
The current version is the one having its version_end_date set to NULL.
If there is not such a version then it means the object has been
'deleted' and so there is no current version available. In this case
the function returns None.
Note that if check_db is False and object's version_end_date is None,
this does not check the database to see if there is a newer version
(perhaps created by some other code), it simply returns the passed
object.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations. See ``VersionManager.version_as_of`` for details on
valid ``relations_as_of`` values.
:param Versionable object: object whose current version will be
returned.
:param mixed relations_as_of: determines point in time used to access
relations. 'start'|'end'|datetime|None
:param bool check_db: Whether or not to look in the database for a
more recent version
:return: Versionable
"""
if object.version_end_date is None and not check_db:
current = object
else:
current = self.current.filter(identity=object.identity).first()
return self.adjust_version_as_of(current, relations_as_of) | [
"def",
"current_version",
"(",
"self",
",",
"object",
",",
"relations_as_of",
"=",
"None",
",",
"check_db",
"=",
"False",
")",
":",
"if",
"object",
".",
"version_end_date",
"is",
"None",
"and",
"not",
"check_db",
":",
"current",
"=",
"object",
"else",
":",
"current",
"=",
"self",
".",
"current",
".",
"filter",
"(",
"identity",
"=",
"object",
".",
"identity",
")",
".",
"first",
"(",
")",
"return",
"self",
".",
"adjust_version_as_of",
"(",
"current",
",",
"relations_as_of",
")"
] | Return the current version of the given object.
The current version is the one having its version_end_date set to NULL.
If there is not such a version then it means the object has been
'deleted' and so there is no current version available. In this case
the function returns None.
Note that if check_db is False and object's version_end_date is None,
this does not check the database to see if there is a newer version
(perhaps created by some other code), it simply returns the passed
object.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations. See ``VersionManager.version_as_of`` for details on
valid ``relations_as_of`` values.
:param Versionable object: object whose current version will be
returned.
:param mixed relations_as_of: determines point in time used to access
relations. 'start'|'end'|datetime|None
:param bool check_db: Whether or not to look in the database for a
more recent version
:return: Versionable | [
"Return",
"the",
"current",
"version",
"of",
"the",
"given",
"object",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L153-L185 | train |
swisscom/cleanerversion | versions/models.py | VersionManager.adjust_version_as_of | def adjust_version_as_of(version, relations_as_of):
"""
Adjusts the passed version's as_of time to an appropriate value, and
returns it.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations.
Valid ``relations_as_of`` values and how this affects the returned
version's as_of attribute:
- 'start': version start date
- 'end': version end date - 1 microsecond (no effect if version is
current version)
- datetime object: given datetime (raises ValueError if given datetime
not valid for version)
- None: unset (related object queries will not be restricted to a
point in time)
:param Versionable object: object whose as_of will be adjusted as
requested.
:param mixed relations_as_of: valid values are the strings 'start' or
'end', or a datetime object.
:return: Versionable
"""
if not version:
return version
if relations_as_of == 'end':
if version.is_current:
# Ensure that version._querytime is active, in case it wasn't
# before.
version.as_of = None
else:
version.as_of = version.version_end_date - datetime.timedelta(
microseconds=1)
elif relations_as_of == 'start':
version.as_of = version.version_start_date
elif isinstance(relations_as_of, datetime.datetime):
as_of = relations_as_of.astimezone(utc)
if not as_of >= version.version_start_date:
raise ValueError(
"Provided as_of '{}' is earlier than version's start "
"time '{}'".format(
as_of.isoformat(),
version.version_start_date.isoformat()
)
)
if version.version_end_date is not None \
and as_of >= version.version_end_date:
raise ValueError(
"Provided as_of '{}' is later than version's start "
"time '{}'".format(
as_of.isoformat(),
version.version_end_date.isoformat()
)
)
version.as_of = as_of
elif relations_as_of is None:
version._querytime = QueryTime(time=None, active=False)
else:
raise TypeError(
"as_of parameter must be 'start', 'end', None, or datetime "
"object")
return version | python | def adjust_version_as_of(version, relations_as_of):
"""
Adjusts the passed version's as_of time to an appropriate value, and
returns it.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations.
Valid ``relations_as_of`` values and how this affects the returned
version's as_of attribute:
- 'start': version start date
- 'end': version end date - 1 microsecond (no effect if version is
current version)
- datetime object: given datetime (raises ValueError if given datetime
not valid for version)
- None: unset (related object queries will not be restricted to a
point in time)
:param Versionable object: object whose as_of will be adjusted as
requested.
:param mixed relations_as_of: valid values are the strings 'start' or
'end', or a datetime object.
:return: Versionable
"""
if not version:
return version
if relations_as_of == 'end':
if version.is_current:
# Ensure that version._querytime is active, in case it wasn't
# before.
version.as_of = None
else:
version.as_of = version.version_end_date - datetime.timedelta(
microseconds=1)
elif relations_as_of == 'start':
version.as_of = version.version_start_date
elif isinstance(relations_as_of, datetime.datetime):
as_of = relations_as_of.astimezone(utc)
if not as_of >= version.version_start_date:
raise ValueError(
"Provided as_of '{}' is earlier than version's start "
"time '{}'".format(
as_of.isoformat(),
version.version_start_date.isoformat()
)
)
if version.version_end_date is not None \
and as_of >= version.version_end_date:
raise ValueError(
"Provided as_of '{}' is later than version's start "
"time '{}'".format(
as_of.isoformat(),
version.version_end_date.isoformat()
)
)
version.as_of = as_of
elif relations_as_of is None:
version._querytime = QueryTime(time=None, active=False)
else:
raise TypeError(
"as_of parameter must be 'start', 'end', None, or datetime "
"object")
return version | [
"def",
"adjust_version_as_of",
"(",
"version",
",",
"relations_as_of",
")",
":",
"if",
"not",
"version",
":",
"return",
"version",
"if",
"relations_as_of",
"==",
"'end'",
":",
"if",
"version",
".",
"is_current",
":",
"# Ensure that version._querytime is active, in case it wasn't",
"# before.",
"version",
".",
"as_of",
"=",
"None",
"else",
":",
"version",
".",
"as_of",
"=",
"version",
".",
"version_end_date",
"-",
"datetime",
".",
"timedelta",
"(",
"microseconds",
"=",
"1",
")",
"elif",
"relations_as_of",
"==",
"'start'",
":",
"version",
".",
"as_of",
"=",
"version",
".",
"version_start_date",
"elif",
"isinstance",
"(",
"relations_as_of",
",",
"datetime",
".",
"datetime",
")",
":",
"as_of",
"=",
"relations_as_of",
".",
"astimezone",
"(",
"utc",
")",
"if",
"not",
"as_of",
">=",
"version",
".",
"version_start_date",
":",
"raise",
"ValueError",
"(",
"\"Provided as_of '{}' is earlier than version's start \"",
"\"time '{}'\"",
".",
"format",
"(",
"as_of",
".",
"isoformat",
"(",
")",
",",
"version",
".",
"version_start_date",
".",
"isoformat",
"(",
")",
")",
")",
"if",
"version",
".",
"version_end_date",
"is",
"not",
"None",
"and",
"as_of",
">=",
"version",
".",
"version_end_date",
":",
"raise",
"ValueError",
"(",
"\"Provided as_of '{}' is later than version's start \"",
"\"time '{}'\"",
".",
"format",
"(",
"as_of",
".",
"isoformat",
"(",
")",
",",
"version",
".",
"version_end_date",
".",
"isoformat",
"(",
")",
")",
")",
"version",
".",
"as_of",
"=",
"as_of",
"elif",
"relations_as_of",
"is",
"None",
":",
"version",
".",
"_querytime",
"=",
"QueryTime",
"(",
"time",
"=",
"None",
",",
"active",
"=",
"False",
")",
"else",
":",
"raise",
"TypeError",
"(",
"\"as_of parameter must be 'start', 'end', None, or datetime \"",
"\"object\"",
")",
"return",
"version"
] | Adjusts the passed version's as_of time to an appropriate value, and
returns it.
``relations_as_of`` is used to fix the point in time for the version;
this affects which related objects are returned when querying for
object relations.
Valid ``relations_as_of`` values and how this affects the returned
version's as_of attribute:
- 'start': version start date
- 'end': version end date - 1 microsecond (no effect if version is
current version)
- datetime object: given datetime (raises ValueError if given datetime
not valid for version)
- None: unset (related object queries will not be restricted to a
point in time)
:param Versionable object: object whose as_of will be adjusted as
requested.
:param mixed relations_as_of: valid values are the strings 'start' or
'end', or a datetime object.
:return: Versionable | [
"Adjusts",
"the",
"passed",
"version",
"s",
"as_of",
"time",
"to",
"an",
"appropriate",
"value",
"and",
"returns",
"it",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L188-L252 | train |
swisscom/cleanerversion | versions/models.py | VersionedQuerySet._fetch_all | def _fetch_all(self):
"""
Completely overrides the QuerySet._fetch_all method by adding the
timestamp to all objects
:return: See django.db.models.query.QuerySet._fetch_all for return
values
"""
if self._result_cache is None:
self._result_cache = list(self.iterator())
# TODO: Do we have to test for ValuesListIterable, ValuesIterable,
# and FlatValuesListIterable here?
if self._iterable_class == ModelIterable:
for x in self._result_cache:
self._set_item_querytime(x)
if self._prefetch_related_lookups and not self._prefetch_done:
self._prefetch_related_objects() | python | def _fetch_all(self):
"""
Completely overrides the QuerySet._fetch_all method by adding the
timestamp to all objects
:return: See django.db.models.query.QuerySet._fetch_all for return
values
"""
if self._result_cache is None:
self._result_cache = list(self.iterator())
# TODO: Do we have to test for ValuesListIterable, ValuesIterable,
# and FlatValuesListIterable here?
if self._iterable_class == ModelIterable:
for x in self._result_cache:
self._set_item_querytime(x)
if self._prefetch_related_lookups and not self._prefetch_done:
self._prefetch_related_objects() | [
"def",
"_fetch_all",
"(",
"self",
")",
":",
"if",
"self",
".",
"_result_cache",
"is",
"None",
":",
"self",
".",
"_result_cache",
"=",
"list",
"(",
"self",
".",
"iterator",
"(",
")",
")",
"# TODO: Do we have to test for ValuesListIterable, ValuesIterable,",
"# and FlatValuesListIterable here?",
"if",
"self",
".",
"_iterable_class",
"==",
"ModelIterable",
":",
"for",
"x",
"in",
"self",
".",
"_result_cache",
":",
"self",
".",
"_set_item_querytime",
"(",
"x",
")",
"if",
"self",
".",
"_prefetch_related_lookups",
"and",
"not",
"self",
".",
"_prefetch_done",
":",
"self",
".",
"_prefetch_related_objects",
"(",
")"
] | Completely overrides the QuerySet._fetch_all method by adding the
timestamp to all objects
:return: See django.db.models.query.QuerySet._fetch_all for return
values | [
"Completely",
"overrides",
"the",
"QuerySet",
".",
"_fetch_all",
"method",
"by",
"adding",
"the",
"timestamp",
"to",
"all",
"objects"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L500-L516 | train |
swisscom/cleanerversion | versions/models.py | VersionedQuerySet._clone | def _clone(self, *args, **kwargs):
"""
Overrides the QuerySet._clone method by adding the cloning of the
VersionedQuerySet's query_time parameter
:param kwargs: Same as the original QuerySet._clone params
:return: Just as QuerySet._clone, this method returns a clone of the
original object
"""
clone = super(VersionedQuerySet, self)._clone(**kwargs)
clone.querytime = self.querytime
return clone | python | def _clone(self, *args, **kwargs):
"""
Overrides the QuerySet._clone method by adding the cloning of the
VersionedQuerySet's query_time parameter
:param kwargs: Same as the original QuerySet._clone params
:return: Just as QuerySet._clone, this method returns a clone of the
original object
"""
clone = super(VersionedQuerySet, self)._clone(**kwargs)
clone.querytime = self.querytime
return clone | [
"def",
"_clone",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"clone",
"=",
"super",
"(",
"VersionedQuerySet",
",",
"self",
")",
".",
"_clone",
"(",
"*",
"*",
"kwargs",
")",
"clone",
".",
"querytime",
"=",
"self",
".",
"querytime",
"return",
"clone"
] | Overrides the QuerySet._clone method by adding the cloning of the
VersionedQuerySet's query_time parameter
:param kwargs: Same as the original QuerySet._clone params
:return: Just as QuerySet._clone, this method returns a clone of the
original object | [
"Overrides",
"the",
"QuerySet",
".",
"_clone",
"method",
"by",
"adding",
"the",
"cloning",
"of",
"the",
"VersionedQuerySet",
"s",
"query_time",
"parameter"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L518-L529 | train |
swisscom/cleanerversion | versions/models.py | VersionedQuerySet._set_item_querytime | def _set_item_querytime(self, item, type_check=True):
"""
Sets the time for which the query was made on the resulting item
:param item: an item of type Versionable
:param type_check: Check the item to be a Versionable
:return: Returns the item itself with the time set
"""
if isinstance(item, Versionable):
item._querytime = self.querytime
elif isinstance(item, VersionedQuerySet):
item.querytime = self.querytime
else:
if type_check:
raise TypeError(
"This item is not a Versionable, it's a " + str(
type(item)))
return item | python | def _set_item_querytime(self, item, type_check=True):
"""
Sets the time for which the query was made on the resulting item
:param item: an item of type Versionable
:param type_check: Check the item to be a Versionable
:return: Returns the item itself with the time set
"""
if isinstance(item, Versionable):
item._querytime = self.querytime
elif isinstance(item, VersionedQuerySet):
item.querytime = self.querytime
else:
if type_check:
raise TypeError(
"This item is not a Versionable, it's a " + str(
type(item)))
return item | [
"def",
"_set_item_querytime",
"(",
"self",
",",
"item",
",",
"type_check",
"=",
"True",
")",
":",
"if",
"isinstance",
"(",
"item",
",",
"Versionable",
")",
":",
"item",
".",
"_querytime",
"=",
"self",
".",
"querytime",
"elif",
"isinstance",
"(",
"item",
",",
"VersionedQuerySet",
")",
":",
"item",
".",
"querytime",
"=",
"self",
".",
"querytime",
"else",
":",
"if",
"type_check",
":",
"raise",
"TypeError",
"(",
"\"This item is not a Versionable, it's a \"",
"+",
"str",
"(",
"type",
"(",
"item",
")",
")",
")",
"return",
"item"
] | Sets the time for which the query was made on the resulting item
:param item: an item of type Versionable
:param type_check: Check the item to be a Versionable
:return: Returns the item itself with the time set | [
"Sets",
"the",
"time",
"for",
"which",
"the",
"query",
"was",
"made",
"on",
"the",
"resulting",
"item"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L531-L548 | train |
swisscom/cleanerversion | versions/models.py | VersionedQuerySet.as_of | def as_of(self, qtime=None):
"""
Sets the time for which we want to retrieve an object.
:param qtime: The UTC date and time; if None then use the current
state (where version_end_date = NULL)
:return: A VersionedQuerySet
"""
clone = self._clone()
clone.querytime = QueryTime(time=qtime, active=True)
return clone | python | def as_of(self, qtime=None):
"""
Sets the time for which we want to retrieve an object.
:param qtime: The UTC date and time; if None then use the current
state (where version_end_date = NULL)
:return: A VersionedQuerySet
"""
clone = self._clone()
clone.querytime = QueryTime(time=qtime, active=True)
return clone | [
"def",
"as_of",
"(",
"self",
",",
"qtime",
"=",
"None",
")",
":",
"clone",
"=",
"self",
".",
"_clone",
"(",
")",
"clone",
".",
"querytime",
"=",
"QueryTime",
"(",
"time",
"=",
"qtime",
",",
"active",
"=",
"True",
")",
"return",
"clone"
] | Sets the time for which we want to retrieve an object.
:param qtime: The UTC date and time; if None then use the current
state (where version_end_date = NULL)
:return: A VersionedQuerySet | [
"Sets",
"the",
"time",
"for",
"which",
"we",
"want",
"to",
"retrieve",
"an",
"object",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L550-L560 | train |
swisscom/cleanerversion | versions/models.py | VersionedQuerySet.delete | def delete(self):
"""
Deletes the records in the QuerySet.
"""
assert self.query.can_filter(), \
"Cannot use 'limit' or 'offset' with delete."
# Ensure that only current objects are selected.
del_query = self.filter(version_end_date__isnull=True)
# The delete is actually 2 queries - one to find related objects,
# and one to delete. Make sure that the discovery of related
# objects is performed on the same database as the deletion.
del_query._for_write = True
# Disable non-supported fields.
del_query.query.select_for_update = False
del_query.query.select_related = False
del_query.query.clear_ordering(force_empty=True)
collector_class = get_versioned_delete_collector_class()
collector = collector_class(using=del_query.db)
collector.collect(del_query)
collector.delete(get_utc_now())
# Clear the result cache, in case this QuerySet gets reused.
self._result_cache = None | python | def delete(self):
"""
Deletes the records in the QuerySet.
"""
assert self.query.can_filter(), \
"Cannot use 'limit' or 'offset' with delete."
# Ensure that only current objects are selected.
del_query = self.filter(version_end_date__isnull=True)
# The delete is actually 2 queries - one to find related objects,
# and one to delete. Make sure that the discovery of related
# objects is performed on the same database as the deletion.
del_query._for_write = True
# Disable non-supported fields.
del_query.query.select_for_update = False
del_query.query.select_related = False
del_query.query.clear_ordering(force_empty=True)
collector_class = get_versioned_delete_collector_class()
collector = collector_class(using=del_query.db)
collector.collect(del_query)
collector.delete(get_utc_now())
# Clear the result cache, in case this QuerySet gets reused.
self._result_cache = None | [
"def",
"delete",
"(",
"self",
")",
":",
"assert",
"self",
".",
"query",
".",
"can_filter",
"(",
")",
",",
"\"Cannot use 'limit' or 'offset' with delete.\"",
"# Ensure that only current objects are selected.",
"del_query",
"=",
"self",
".",
"filter",
"(",
"version_end_date__isnull",
"=",
"True",
")",
"# The delete is actually 2 queries - one to find related objects,",
"# and one to delete. Make sure that the discovery of related",
"# objects is performed on the same database as the deletion.",
"del_query",
".",
"_for_write",
"=",
"True",
"# Disable non-supported fields.",
"del_query",
".",
"query",
".",
"select_for_update",
"=",
"False",
"del_query",
".",
"query",
".",
"select_related",
"=",
"False",
"del_query",
".",
"query",
".",
"clear_ordering",
"(",
"force_empty",
"=",
"True",
")",
"collector_class",
"=",
"get_versioned_delete_collector_class",
"(",
")",
"collector",
"=",
"collector_class",
"(",
"using",
"=",
"del_query",
".",
"db",
")",
"collector",
".",
"collect",
"(",
"del_query",
")",
"collector",
".",
"delete",
"(",
"get_utc_now",
"(",
")",
")",
"# Clear the result cache, in case this QuerySet gets reused.",
"self",
".",
"_result_cache",
"=",
"None"
] | Deletes the records in the QuerySet. | [
"Deletes",
"the",
"records",
"in",
"the",
"QuerySet",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L562-L588 | train |
swisscom/cleanerversion | versions/models.py | Versionable.uuid | def uuid(uuid_value=None):
"""
Returns a uuid value that is valid to use for id and identity fields.
:return: unicode uuid object if using UUIDFields, uuid unicode string
otherwise.
"""
if uuid_value:
if not validate_uuid(uuid_value):
raise ValueError(
"uuid_value must be a valid UUID version 4 object")
else:
uuid_value = uuid.uuid4()
if versions_settings.VERSIONS_USE_UUIDFIELD:
return uuid_value
else:
return six.u(str(uuid_value)) | python | def uuid(uuid_value=None):
"""
Returns a uuid value that is valid to use for id and identity fields.
:return: unicode uuid object if using UUIDFields, uuid unicode string
otherwise.
"""
if uuid_value:
if not validate_uuid(uuid_value):
raise ValueError(
"uuid_value must be a valid UUID version 4 object")
else:
uuid_value = uuid.uuid4()
if versions_settings.VERSIONS_USE_UUIDFIELD:
return uuid_value
else:
return six.u(str(uuid_value)) | [
"def",
"uuid",
"(",
"uuid_value",
"=",
"None",
")",
":",
"if",
"uuid_value",
":",
"if",
"not",
"validate_uuid",
"(",
"uuid_value",
")",
":",
"raise",
"ValueError",
"(",
"\"uuid_value must be a valid UUID version 4 object\"",
")",
"else",
":",
"uuid_value",
"=",
"uuid",
".",
"uuid4",
"(",
")",
"if",
"versions_settings",
".",
"VERSIONS_USE_UUIDFIELD",
":",
"return",
"uuid_value",
"else",
":",
"return",
"six",
".",
"u",
"(",
"str",
"(",
"uuid_value",
")",
")"
] | Returns a uuid value that is valid to use for id and identity fields.
:return: unicode uuid object if using UUIDFields, uuid unicode string
otherwise. | [
"Returns",
"a",
"uuid",
"value",
"that",
"is",
"valid",
"to",
"use",
"for",
"id",
"and",
"identity",
"fields",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L737-L754 | train |
swisscom/cleanerversion | versions/models.py | Versionable.restore | def restore(self, **kwargs):
"""
Restores this version as a new version, and returns this new version.
If a current version already exists, it will be terminated before
restoring this version.
Relations (foreign key, reverse foreign key, many-to-many) are not
restored with the old version. If provided in kwargs,
(Versioned)ForeignKey fields will be set to the provided values.
If passing an id for a (Versioned)ForeignKey, use the field.attname.
For example:
restore(team_id=myteam.pk)
If passing an object, simply use the field name, e.g.:
restore(team=myteam)
If a (Versioned)ForeignKey is not nullable and no value is provided
for it in kwargs, a ForeignKeyRequiresValueError will be raised.
:param kwargs: arguments used to initialize the class instance
:return: Versionable
"""
if not self.pk:
raise ValueError(
'Instance must be saved and terminated before it can be '
'restored.')
if self.is_current:
raise ValueError(
'This is the current version, no need to restore it.')
if self.get_deferred_fields():
# It would be necessary to fetch the record from the database
# again for this to succeed.
# Alternatively, perhaps it would be possible to create a copy
# of the object after fetching the missing fields.
# Doing so may be unexpected by the calling code, so raise an
# exception: the calling code should be adapted if necessary.
raise ValueError(
'Can not restore a model instance that has deferred fields')
cls = self.__class__
now = get_utc_now()
restored = copy.copy(self)
restored.version_end_date = None
restored.version_start_date = now
fields = [f for f in cls._meta.local_fields if
f.name not in Versionable.VERSIONABLE_FIELDS]
for field in fields:
if field.attname in kwargs:
# Fake an object in order to avoid a DB roundtrip
# This was made necessary, since assigning to the field's
# attname did not work anymore with Django 2.0
obj = field.remote_field.model(id=kwargs[field.attname])
setattr(restored, field.name, obj)
elif field.name in kwargs:
setattr(restored, field.name, kwargs[field.name])
elif isinstance(field, ForeignKey):
# Set all non-provided ForeignKeys to None. If required,
# raise an error.
try:
setattr(restored, field.name, None)
# Check for non null foreign key removed since Django 1.10
# https://docs.djangoproject.com/en/1.10/releases/1.10/
# #removed-null-assignment-check-for-non-null-foreign-
# key-fields
if not field.null:
raise ValueError
except ValueError:
raise ForeignKeyRequiresValueError
self.id = self.uuid()
with transaction.atomic():
# If this is not the latest version, terminate the latest version
latest = cls.objects.current_version(self, check_db=True)
if latest and latest != self:
latest.delete()
restored.version_start_date = latest.version_end_date
self.save()
restored.save()
# Update ManyToMany relations to point to the old version's id
# instead of the restored version's id.
for field_name in self.get_all_m2m_field_names():
manager = getattr(restored,
field_name)
# returns a VersionedRelatedManager instance
manager.through.objects.filter(
**{manager.source_field.attname: restored.id}).update(
**{manager.source_field_name: self})
return restored | python | def restore(self, **kwargs):
"""
Restores this version as a new version, and returns this new version.
If a current version already exists, it will be terminated before
restoring this version.
Relations (foreign key, reverse foreign key, many-to-many) are not
restored with the old version. If provided in kwargs,
(Versioned)ForeignKey fields will be set to the provided values.
If passing an id for a (Versioned)ForeignKey, use the field.attname.
For example:
restore(team_id=myteam.pk)
If passing an object, simply use the field name, e.g.:
restore(team=myteam)
If a (Versioned)ForeignKey is not nullable and no value is provided
for it in kwargs, a ForeignKeyRequiresValueError will be raised.
:param kwargs: arguments used to initialize the class instance
:return: Versionable
"""
if not self.pk:
raise ValueError(
'Instance must be saved and terminated before it can be '
'restored.')
if self.is_current:
raise ValueError(
'This is the current version, no need to restore it.')
if self.get_deferred_fields():
# It would be necessary to fetch the record from the database
# again for this to succeed.
# Alternatively, perhaps it would be possible to create a copy
# of the object after fetching the missing fields.
# Doing so may be unexpected by the calling code, so raise an
# exception: the calling code should be adapted if necessary.
raise ValueError(
'Can not restore a model instance that has deferred fields')
cls = self.__class__
now = get_utc_now()
restored = copy.copy(self)
restored.version_end_date = None
restored.version_start_date = now
fields = [f for f in cls._meta.local_fields if
f.name not in Versionable.VERSIONABLE_FIELDS]
for field in fields:
if field.attname in kwargs:
# Fake an object in order to avoid a DB roundtrip
# This was made necessary, since assigning to the field's
# attname did not work anymore with Django 2.0
obj = field.remote_field.model(id=kwargs[field.attname])
setattr(restored, field.name, obj)
elif field.name in kwargs:
setattr(restored, field.name, kwargs[field.name])
elif isinstance(field, ForeignKey):
# Set all non-provided ForeignKeys to None. If required,
# raise an error.
try:
setattr(restored, field.name, None)
# Check for non null foreign key removed since Django 1.10
# https://docs.djangoproject.com/en/1.10/releases/1.10/
# #removed-null-assignment-check-for-non-null-foreign-
# key-fields
if not field.null:
raise ValueError
except ValueError:
raise ForeignKeyRequiresValueError
self.id = self.uuid()
with transaction.atomic():
# If this is not the latest version, terminate the latest version
latest = cls.objects.current_version(self, check_db=True)
if latest and latest != self:
latest.delete()
restored.version_start_date = latest.version_end_date
self.save()
restored.save()
# Update ManyToMany relations to point to the old version's id
# instead of the restored version's id.
for field_name in self.get_all_m2m_field_names():
manager = getattr(restored,
field_name)
# returns a VersionedRelatedManager instance
manager.through.objects.filter(
**{manager.source_field.attname: restored.id}).update(
**{manager.source_field_name: self})
return restored | [
"def",
"restore",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"if",
"not",
"self",
".",
"pk",
":",
"raise",
"ValueError",
"(",
"'Instance must be saved and terminated before it can be '",
"'restored.'",
")",
"if",
"self",
".",
"is_current",
":",
"raise",
"ValueError",
"(",
"'This is the current version, no need to restore it.'",
")",
"if",
"self",
".",
"get_deferred_fields",
"(",
")",
":",
"# It would be necessary to fetch the record from the database",
"# again for this to succeed.",
"# Alternatively, perhaps it would be possible to create a copy",
"# of the object after fetching the missing fields.",
"# Doing so may be unexpected by the calling code, so raise an",
"# exception: the calling code should be adapted if necessary.",
"raise",
"ValueError",
"(",
"'Can not restore a model instance that has deferred fields'",
")",
"cls",
"=",
"self",
".",
"__class__",
"now",
"=",
"get_utc_now",
"(",
")",
"restored",
"=",
"copy",
".",
"copy",
"(",
"self",
")",
"restored",
".",
"version_end_date",
"=",
"None",
"restored",
".",
"version_start_date",
"=",
"now",
"fields",
"=",
"[",
"f",
"for",
"f",
"in",
"cls",
".",
"_meta",
".",
"local_fields",
"if",
"f",
".",
"name",
"not",
"in",
"Versionable",
".",
"VERSIONABLE_FIELDS",
"]",
"for",
"field",
"in",
"fields",
":",
"if",
"field",
".",
"attname",
"in",
"kwargs",
":",
"# Fake an object in order to avoid a DB roundtrip",
"# This was made necessary, since assigning to the field's",
"# attname did not work anymore with Django 2.0",
"obj",
"=",
"field",
".",
"remote_field",
".",
"model",
"(",
"id",
"=",
"kwargs",
"[",
"field",
".",
"attname",
"]",
")",
"setattr",
"(",
"restored",
",",
"field",
".",
"name",
",",
"obj",
")",
"elif",
"field",
".",
"name",
"in",
"kwargs",
":",
"setattr",
"(",
"restored",
",",
"field",
".",
"name",
",",
"kwargs",
"[",
"field",
".",
"name",
"]",
")",
"elif",
"isinstance",
"(",
"field",
",",
"ForeignKey",
")",
":",
"# Set all non-provided ForeignKeys to None. If required,",
"# raise an error.",
"try",
":",
"setattr",
"(",
"restored",
",",
"field",
".",
"name",
",",
"None",
")",
"# Check for non null foreign key removed since Django 1.10",
"# https://docs.djangoproject.com/en/1.10/releases/1.10/",
"# #removed-null-assignment-check-for-non-null-foreign-",
"# key-fields",
"if",
"not",
"field",
".",
"null",
":",
"raise",
"ValueError",
"except",
"ValueError",
":",
"raise",
"ForeignKeyRequiresValueError",
"self",
".",
"id",
"=",
"self",
".",
"uuid",
"(",
")",
"with",
"transaction",
".",
"atomic",
"(",
")",
":",
"# If this is not the latest version, terminate the latest version",
"latest",
"=",
"cls",
".",
"objects",
".",
"current_version",
"(",
"self",
",",
"check_db",
"=",
"True",
")",
"if",
"latest",
"and",
"latest",
"!=",
"self",
":",
"latest",
".",
"delete",
"(",
")",
"restored",
".",
"version_start_date",
"=",
"latest",
".",
"version_end_date",
"self",
".",
"save",
"(",
")",
"restored",
".",
"save",
"(",
")",
"# Update ManyToMany relations to point to the old version's id",
"# instead of the restored version's id.",
"for",
"field_name",
"in",
"self",
".",
"get_all_m2m_field_names",
"(",
")",
":",
"manager",
"=",
"getattr",
"(",
"restored",
",",
"field_name",
")",
"# returns a VersionedRelatedManager instance",
"manager",
".",
"through",
".",
"objects",
".",
"filter",
"(",
"*",
"*",
"{",
"manager",
".",
"source_field",
".",
"attname",
":",
"restored",
".",
"id",
"}",
")",
".",
"update",
"(",
"*",
"*",
"{",
"manager",
".",
"source_field_name",
":",
"self",
"}",
")",
"return",
"restored"
] | Restores this version as a new version, and returns this new version.
If a current version already exists, it will be terminated before
restoring this version.
Relations (foreign key, reverse foreign key, many-to-many) are not
restored with the old version. If provided in kwargs,
(Versioned)ForeignKey fields will be set to the provided values.
If passing an id for a (Versioned)ForeignKey, use the field.attname.
For example:
restore(team_id=myteam.pk)
If passing an object, simply use the field name, e.g.:
restore(team=myteam)
If a (Versioned)ForeignKey is not nullable and no value is provided
for it in kwargs, a ForeignKeyRequiresValueError will be raised.
:param kwargs: arguments used to initialize the class instance
:return: Versionable | [
"Restores",
"this",
"version",
"as",
"a",
"new",
"version",
"and",
"returns",
"this",
"new",
"version",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L908-L1002 | train |
swisscom/cleanerversion | versions/models.py | Versionable.detach | def detach(self):
"""
Detaches the instance from its history.
Similar to creating a new object with the same field values. The id and
identity fields are set to a new value. The returned object has not
been saved, call save() afterwards when you are ready to persist the
object.
ManyToMany and reverse ForeignKey relations are lost for the detached
object.
:return: Versionable
"""
self.id = self.identity = self.uuid()
self.version_start_date = self.version_birth_date = get_utc_now()
self.version_end_date = None
return self | python | def detach(self):
"""
Detaches the instance from its history.
Similar to creating a new object with the same field values. The id and
identity fields are set to a new value. The returned object has not
been saved, call save() afterwards when you are ready to persist the
object.
ManyToMany and reverse ForeignKey relations are lost for the detached
object.
:return: Versionable
"""
self.id = self.identity = self.uuid()
self.version_start_date = self.version_birth_date = get_utc_now()
self.version_end_date = None
return self | [
"def",
"detach",
"(",
"self",
")",
":",
"self",
".",
"id",
"=",
"self",
".",
"identity",
"=",
"self",
".",
"uuid",
"(",
")",
"self",
".",
"version_start_date",
"=",
"self",
".",
"version_birth_date",
"=",
"get_utc_now",
"(",
")",
"self",
".",
"version_end_date",
"=",
"None",
"return",
"self"
] | Detaches the instance from its history.
Similar to creating a new object with the same field values. The id and
identity fields are set to a new value. The returned object has not
been saved, call save() afterwards when you are ready to persist the
object.
ManyToMany and reverse ForeignKey relations are lost for the detached
object.
:return: Versionable | [
"Detaches",
"the",
"instance",
"from",
"its",
"history",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L1013-L1030 | train |
swisscom/cleanerversion | versions/models.py | Versionable.matches_querytime | def matches_querytime(instance, querytime):
"""
Checks whether the given instance satisfies the given QueryTime object.
:param instance: an instance of Versionable
:param querytime: QueryTime value to check against
"""
if not querytime.active:
return True
if not querytime.time:
return instance.version_end_date is None
return (instance.version_start_date <= querytime.time and
(instance.version_end_date is None or
instance.version_end_date > querytime.time)) | python | def matches_querytime(instance, querytime):
"""
Checks whether the given instance satisfies the given QueryTime object.
:param instance: an instance of Versionable
:param querytime: QueryTime value to check against
"""
if not querytime.active:
return True
if not querytime.time:
return instance.version_end_date is None
return (instance.version_start_date <= querytime.time and
(instance.version_end_date is None or
instance.version_end_date > querytime.time)) | [
"def",
"matches_querytime",
"(",
"instance",
",",
"querytime",
")",
":",
"if",
"not",
"querytime",
".",
"active",
":",
"return",
"True",
"if",
"not",
"querytime",
".",
"time",
":",
"return",
"instance",
".",
"version_end_date",
"is",
"None",
"return",
"(",
"instance",
".",
"version_start_date",
"<=",
"querytime",
".",
"time",
"and",
"(",
"instance",
".",
"version_end_date",
"is",
"None",
"or",
"instance",
".",
"version_end_date",
">",
"querytime",
".",
"time",
")",
")"
] | Checks whether the given instance satisfies the given QueryTime object.
:param instance: an instance of Versionable
:param querytime: QueryTime value to check against | [
"Checks",
"whether",
"the",
"given",
"instance",
"satisfies",
"the",
"given",
"QueryTime",
"object",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/models.py#L1033-L1048 | train |
swisscom/cleanerversion | versions/fields.py | VersionedForeignKey.contribute_to_related_class | def contribute_to_related_class(self, cls, related):
"""
Override ForeignKey's methods, and replace the descriptor, if set by
the parent's methods
"""
# Internal FK's - i.e., those with a related name ending with '+' -
# and swapped models don't get a related descriptor.
super(VersionedForeignKey, self).contribute_to_related_class(cls,
related)
accessor_name = related.get_accessor_name()
if hasattr(cls, accessor_name):
setattr(cls, accessor_name,
VersionedReverseManyToOneDescriptor(related)) | python | def contribute_to_related_class(self, cls, related):
"""
Override ForeignKey's methods, and replace the descriptor, if set by
the parent's methods
"""
# Internal FK's - i.e., those with a related name ending with '+' -
# and swapped models don't get a related descriptor.
super(VersionedForeignKey, self).contribute_to_related_class(cls,
related)
accessor_name = related.get_accessor_name()
if hasattr(cls, accessor_name):
setattr(cls, accessor_name,
VersionedReverseManyToOneDescriptor(related)) | [
"def",
"contribute_to_related_class",
"(",
"self",
",",
"cls",
",",
"related",
")",
":",
"# Internal FK's - i.e., those with a related name ending with '+' -",
"# and swapped models don't get a related descriptor.",
"super",
"(",
"VersionedForeignKey",
",",
"self",
")",
".",
"contribute_to_related_class",
"(",
"cls",
",",
"related",
")",
"accessor_name",
"=",
"related",
".",
"get_accessor_name",
"(",
")",
"if",
"hasattr",
"(",
"cls",
",",
"accessor_name",
")",
":",
"setattr",
"(",
"cls",
",",
"accessor_name",
",",
"VersionedReverseManyToOneDescriptor",
"(",
"related",
")",
")"
] | Override ForeignKey's methods, and replace the descriptor, if set by
the parent's methods | [
"Override",
"ForeignKey",
"s",
"methods",
"and",
"replace",
"the",
"descriptor",
"if",
"set",
"by",
"the",
"parent",
"s",
"methods"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/fields.py#L32-L44 | train |
swisscom/cleanerversion | versions/fields.py | VersionedForeignKey.get_joining_columns | def get_joining_columns(self, reverse_join=False):
"""
Get and return joining columns defined by this foreign key relationship
:return: A tuple containing the column names of the tables to be
joined (<local_col_name>, <remote_col_name>)
:rtype: tuple
"""
source = self.reverse_related_fields if reverse_join \
else self.related_fields
joining_columns = tuple()
for lhs_field, rhs_field in source:
lhs_col_name = lhs_field.column
rhs_col_name = rhs_field.column
# Test whether
# - self is the current ForeignKey relationship
# - self was not auto_created (e.g. is not part of a M2M
# relationship)
if self is lhs_field and not self.auto_created:
if rhs_col_name == Versionable.VERSION_IDENTIFIER_FIELD:
rhs_col_name = Versionable.OBJECT_IDENTIFIER_FIELD
elif self is rhs_field and not self.auto_created:
if lhs_col_name == Versionable.VERSION_IDENTIFIER_FIELD:
lhs_col_name = Versionable.OBJECT_IDENTIFIER_FIELD
joining_columns = joining_columns + ((lhs_col_name, rhs_col_name),)
return joining_columns | python | def get_joining_columns(self, reverse_join=False):
"""
Get and return joining columns defined by this foreign key relationship
:return: A tuple containing the column names of the tables to be
joined (<local_col_name>, <remote_col_name>)
:rtype: tuple
"""
source = self.reverse_related_fields if reverse_join \
else self.related_fields
joining_columns = tuple()
for lhs_field, rhs_field in source:
lhs_col_name = lhs_field.column
rhs_col_name = rhs_field.column
# Test whether
# - self is the current ForeignKey relationship
# - self was not auto_created (e.g. is not part of a M2M
# relationship)
if self is lhs_field and not self.auto_created:
if rhs_col_name == Versionable.VERSION_IDENTIFIER_FIELD:
rhs_col_name = Versionable.OBJECT_IDENTIFIER_FIELD
elif self is rhs_field and not self.auto_created:
if lhs_col_name == Versionable.VERSION_IDENTIFIER_FIELD:
lhs_col_name = Versionable.OBJECT_IDENTIFIER_FIELD
joining_columns = joining_columns + ((lhs_col_name, rhs_col_name),)
return joining_columns | [
"def",
"get_joining_columns",
"(",
"self",
",",
"reverse_join",
"=",
"False",
")",
":",
"source",
"=",
"self",
".",
"reverse_related_fields",
"if",
"reverse_join",
"else",
"self",
".",
"related_fields",
"joining_columns",
"=",
"tuple",
"(",
")",
"for",
"lhs_field",
",",
"rhs_field",
"in",
"source",
":",
"lhs_col_name",
"=",
"lhs_field",
".",
"column",
"rhs_col_name",
"=",
"rhs_field",
".",
"column",
"# Test whether",
"# - self is the current ForeignKey relationship",
"# - self was not auto_created (e.g. is not part of a M2M",
"# relationship)",
"if",
"self",
"is",
"lhs_field",
"and",
"not",
"self",
".",
"auto_created",
":",
"if",
"rhs_col_name",
"==",
"Versionable",
".",
"VERSION_IDENTIFIER_FIELD",
":",
"rhs_col_name",
"=",
"Versionable",
".",
"OBJECT_IDENTIFIER_FIELD",
"elif",
"self",
"is",
"rhs_field",
"and",
"not",
"self",
".",
"auto_created",
":",
"if",
"lhs_col_name",
"==",
"Versionable",
".",
"VERSION_IDENTIFIER_FIELD",
":",
"lhs_col_name",
"=",
"Versionable",
".",
"OBJECT_IDENTIFIER_FIELD",
"joining_columns",
"=",
"joining_columns",
"+",
"(",
"(",
"lhs_col_name",
",",
"rhs_col_name",
")",
",",
")",
"return",
"joining_columns"
] | Get and return joining columns defined by this foreign key relationship
:return: A tuple containing the column names of the tables to be
joined (<local_col_name>, <remote_col_name>)
:rtype: tuple | [
"Get",
"and",
"return",
"joining",
"columns",
"defined",
"by",
"this",
"foreign",
"key",
"relationship"
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/fields.py#L65-L90 | train |
swisscom/cleanerversion | versions/settings.py | get_versioned_delete_collector_class | def get_versioned_delete_collector_class():
"""
Gets the class to use for deletion collection.
:return: class
"""
key = 'VERSIONED_DELETE_COLLECTOR'
try:
cls = _cache[key]
except KeyError:
collector_class_string = getattr(settings, key)
cls = import_from_string(collector_class_string, key)
_cache[key] = cls
return cls | python | def get_versioned_delete_collector_class():
"""
Gets the class to use for deletion collection.
:return: class
"""
key = 'VERSIONED_DELETE_COLLECTOR'
try:
cls = _cache[key]
except KeyError:
collector_class_string = getattr(settings, key)
cls = import_from_string(collector_class_string, key)
_cache[key] = cls
return cls | [
"def",
"get_versioned_delete_collector_class",
"(",
")",
":",
"key",
"=",
"'VERSIONED_DELETE_COLLECTOR'",
"try",
":",
"cls",
"=",
"_cache",
"[",
"key",
"]",
"except",
"KeyError",
":",
"collector_class_string",
"=",
"getattr",
"(",
"settings",
",",
"key",
")",
"cls",
"=",
"import_from_string",
"(",
"collector_class_string",
",",
"key",
")",
"_cache",
"[",
"key",
"]",
"=",
"cls",
"return",
"cls"
] | Gets the class to use for deletion collection.
:return: class | [
"Gets",
"the",
"class",
"to",
"use",
"for",
"deletion",
"collection",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/settings.py#L57-L70 | train |
swisscom/cleanerversion | versions/deletion.py | VersionedCollector.related_objects | def related_objects(self, related, objs):
"""
Gets a QuerySet of current objects related to ``objs`` via the
relation ``related``.
"""
from versions.models import Versionable
related_model = related.related_model
if issubclass(related_model, Versionable):
qs = related_model.objects.current
else:
qs = related_model._base_manager.all()
return qs.using(self.using).filter(
**{"%s__in" % related.field.name: objs}
) | python | def related_objects(self, related, objs):
"""
Gets a QuerySet of current objects related to ``objs`` via the
relation ``related``.
"""
from versions.models import Versionable
related_model = related.related_model
if issubclass(related_model, Versionable):
qs = related_model.objects.current
else:
qs = related_model._base_manager.all()
return qs.using(self.using).filter(
**{"%s__in" % related.field.name: objs}
) | [
"def",
"related_objects",
"(",
"self",
",",
"related",
",",
"objs",
")",
":",
"from",
"versions",
".",
"models",
"import",
"Versionable",
"related_model",
"=",
"related",
".",
"related_model",
"if",
"issubclass",
"(",
"related_model",
",",
"Versionable",
")",
":",
"qs",
"=",
"related_model",
".",
"objects",
".",
"current",
"else",
":",
"qs",
"=",
"related_model",
".",
"_base_manager",
".",
"all",
"(",
")",
"return",
"qs",
".",
"using",
"(",
"self",
".",
"using",
")",
".",
"filter",
"(",
"*",
"*",
"{",
"\"%s__in\"",
"%",
"related",
".",
"field",
".",
"name",
":",
"objs",
"}",
")"
] | Gets a QuerySet of current objects related to ``objs`` via the
relation ``related``. | [
"Gets",
"a",
"QuerySet",
"of",
"current",
"objects",
"related",
"to",
"objs",
"via",
"the",
"relation",
"related",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/deletion.py#L149-L163 | train |
swisscom/cleanerversion | versions/deletion.py | VersionedCollector.versionable_delete | def versionable_delete(self, instance, timestamp):
"""
Soft-deletes the instance, setting it's version_end_date to timestamp.
Override this method to implement custom behaviour.
:param Versionable instance:
:param datetime timestamp:
"""
instance._delete_at(timestamp, using=self.using) | python | def versionable_delete(self, instance, timestamp):
"""
Soft-deletes the instance, setting it's version_end_date to timestamp.
Override this method to implement custom behaviour.
:param Versionable instance:
:param datetime timestamp:
"""
instance._delete_at(timestamp, using=self.using) | [
"def",
"versionable_delete",
"(",
"self",
",",
"instance",
",",
"timestamp",
")",
":",
"instance",
".",
"_delete_at",
"(",
"timestamp",
",",
"using",
"=",
"self",
".",
"using",
")"
] | Soft-deletes the instance, setting it's version_end_date to timestamp.
Override this method to implement custom behaviour.
:param Versionable instance:
:param datetime timestamp: | [
"Soft",
"-",
"deletes",
"the",
"instance",
"setting",
"it",
"s",
"version_end_date",
"to",
"timestamp",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/deletion.py#L185-L194 | train |
swisscom/cleanerversion | versions/descriptors.py | VersionedManyToManyDescriptor.pks_from_objects | def pks_from_objects(self, objects):
"""
Extract all the primary key strings from the given objects.
Objects may be Versionables, or bare primary keys.
:rtype : set
"""
return {o.pk if isinstance(o, Model) else o for o in objects} | python | def pks_from_objects(self, objects):
"""
Extract all the primary key strings from the given objects.
Objects may be Versionables, or bare primary keys.
:rtype : set
"""
return {o.pk if isinstance(o, Model) else o for o in objects} | [
"def",
"pks_from_objects",
"(",
"self",
",",
"objects",
")",
":",
"return",
"{",
"o",
".",
"pk",
"if",
"isinstance",
"(",
"o",
",",
"Model",
")",
"else",
"o",
"for",
"o",
"in",
"objects",
"}"
] | Extract all the primary key strings from the given objects.
Objects may be Versionables, or bare primary keys.
:rtype : set | [
"Extract",
"all",
"the",
"primary",
"key",
"strings",
"from",
"the",
"given",
"objects",
".",
"Objects",
"may",
"be",
"Versionables",
"or",
"bare",
"primary",
"keys",
"."
] | becadbab5d7b474a0e9a596b99e97682402d2f2c | https://github.com/swisscom/cleanerversion/blob/becadbab5d7b474a0e9a596b99e97682402d2f2c/versions/descriptors.py#L395-L402 | train |
matsui528/nanopq | nanopq/pq.py | PQ.fit | def fit(self, vecs, iter=20, seed=123):
"""Given training vectors, run k-means for each sub-space and create
codewords for each sub-space.
This function should be run once first of all.
Args:
vecs (np.ndarray): Training vectors with shape=(N, D) and dtype=np.float32.
iter (int): The number of iteration for k-means
seed (int): The seed for random process
Returns:
object: self
"""
assert vecs.dtype == np.float32
assert vecs.ndim == 2
N, D = vecs.shape
assert self.Ks < N, "the number of training vector should be more than Ks"
assert D % self.M == 0, "input dimension must be dividable by M"
self.Ds = int(D / self.M)
np.random.seed(seed)
if self.verbose:
print("iter: {}, seed: {}".format(iter, seed))
# [m][ks][ds]: m-th subspace, ks-the codeword, ds-th dim
self.codewords = np.zeros((self.M, self.Ks, self.Ds), dtype=np.float32)
for m in range(self.M):
if self.verbose:
print("Training the subspace: {} / {}".format(m, self.M))
vecs_sub = vecs[:, m * self.Ds : (m+1) * self.Ds]
self.codewords[m], _ = kmeans2(vecs_sub, self.Ks, iter=iter, minit='points')
return self | python | def fit(self, vecs, iter=20, seed=123):
"""Given training vectors, run k-means for each sub-space and create
codewords for each sub-space.
This function should be run once first of all.
Args:
vecs (np.ndarray): Training vectors with shape=(N, D) and dtype=np.float32.
iter (int): The number of iteration for k-means
seed (int): The seed for random process
Returns:
object: self
"""
assert vecs.dtype == np.float32
assert vecs.ndim == 2
N, D = vecs.shape
assert self.Ks < N, "the number of training vector should be more than Ks"
assert D % self.M == 0, "input dimension must be dividable by M"
self.Ds = int(D / self.M)
np.random.seed(seed)
if self.verbose:
print("iter: {}, seed: {}".format(iter, seed))
# [m][ks][ds]: m-th subspace, ks-the codeword, ds-th dim
self.codewords = np.zeros((self.M, self.Ks, self.Ds), dtype=np.float32)
for m in range(self.M):
if self.verbose:
print("Training the subspace: {} / {}".format(m, self.M))
vecs_sub = vecs[:, m * self.Ds : (m+1) * self.Ds]
self.codewords[m], _ = kmeans2(vecs_sub, self.Ks, iter=iter, minit='points')
return self | [
"def",
"fit",
"(",
"self",
",",
"vecs",
",",
"iter",
"=",
"20",
",",
"seed",
"=",
"123",
")",
":",
"assert",
"vecs",
".",
"dtype",
"==",
"np",
".",
"float32",
"assert",
"vecs",
".",
"ndim",
"==",
"2",
"N",
",",
"D",
"=",
"vecs",
".",
"shape",
"assert",
"self",
".",
"Ks",
"<",
"N",
",",
"\"the number of training vector should be more than Ks\"",
"assert",
"D",
"%",
"self",
".",
"M",
"==",
"0",
",",
"\"input dimension must be dividable by M\"",
"self",
".",
"Ds",
"=",
"int",
"(",
"D",
"/",
"self",
".",
"M",
")",
"np",
".",
"random",
".",
"seed",
"(",
"seed",
")",
"if",
"self",
".",
"verbose",
":",
"print",
"(",
"\"iter: {}, seed: {}\"",
".",
"format",
"(",
"iter",
",",
"seed",
")",
")",
"# [m][ks][ds]: m-th subspace, ks-the codeword, ds-th dim",
"self",
".",
"codewords",
"=",
"np",
".",
"zeros",
"(",
"(",
"self",
".",
"M",
",",
"self",
".",
"Ks",
",",
"self",
".",
"Ds",
")",
",",
"dtype",
"=",
"np",
".",
"float32",
")",
"for",
"m",
"in",
"range",
"(",
"self",
".",
"M",
")",
":",
"if",
"self",
".",
"verbose",
":",
"print",
"(",
"\"Training the subspace: {} / {}\"",
".",
"format",
"(",
"m",
",",
"self",
".",
"M",
")",
")",
"vecs_sub",
"=",
"vecs",
"[",
":",
",",
"m",
"*",
"self",
".",
"Ds",
":",
"(",
"m",
"+",
"1",
")",
"*",
"self",
".",
"Ds",
"]",
"self",
".",
"codewords",
"[",
"m",
"]",
",",
"_",
"=",
"kmeans2",
"(",
"vecs_sub",
",",
"self",
".",
"Ks",
",",
"iter",
"=",
"iter",
",",
"minit",
"=",
"'points'",
")",
"return",
"self"
] | Given training vectors, run k-means for each sub-space and create
codewords for each sub-space.
This function should be run once first of all.
Args:
vecs (np.ndarray): Training vectors with shape=(N, D) and dtype=np.float32.
iter (int): The number of iteration for k-means
seed (int): The seed for random process
Returns:
object: self | [
"Given",
"training",
"vectors",
"run",
"k",
"-",
"means",
"for",
"each",
"sub",
"-",
"space",
"and",
"create",
"codewords",
"for",
"each",
"sub",
"-",
"space",
"."
] | 1ce68cad2e3cab62b409e6dd63f676ed7b443ee9 | https://github.com/matsui528/nanopq/blob/1ce68cad2e3cab62b409e6dd63f676ed7b443ee9/nanopq/pq.py#L53-L87 | train |
matsui528/nanopq | nanopq/pq.py | PQ.encode | def encode(self, vecs):
"""Encode input vectors into PQ-codes.
Args:
vecs (np.ndarray): Input vectors with shape=(N, D) and dtype=np.float32.
Returns:
np.ndarray: PQ codes with shape=(N, M) and dtype=self.code_dtype
"""
assert vecs.dtype == np.float32
assert vecs.ndim == 2
N, D = vecs.shape
assert D == self.Ds * self.M, "input dimension must be Ds * M"
# codes[n][m] : code of n-th vec, m-th subspace
codes = np.empty((N, self.M), dtype=self.code_dtype)
for m in range(self.M):
if self.verbose:
print("Encoding the subspace: {} / {}".format(m, self.M))
vecs_sub = vecs[:, m * self.Ds : (m+1) * self.Ds]
codes[:, m], _ = vq(vecs_sub, self.codewords[m])
return codes | python | def encode(self, vecs):
"""Encode input vectors into PQ-codes.
Args:
vecs (np.ndarray): Input vectors with shape=(N, D) and dtype=np.float32.
Returns:
np.ndarray: PQ codes with shape=(N, M) and dtype=self.code_dtype
"""
assert vecs.dtype == np.float32
assert vecs.ndim == 2
N, D = vecs.shape
assert D == self.Ds * self.M, "input dimension must be Ds * M"
# codes[n][m] : code of n-th vec, m-th subspace
codes = np.empty((N, self.M), dtype=self.code_dtype)
for m in range(self.M):
if self.verbose:
print("Encoding the subspace: {} / {}".format(m, self.M))
vecs_sub = vecs[:, m * self.Ds : (m+1) * self.Ds]
codes[:, m], _ = vq(vecs_sub, self.codewords[m])
return codes | [
"def",
"encode",
"(",
"self",
",",
"vecs",
")",
":",
"assert",
"vecs",
".",
"dtype",
"==",
"np",
".",
"float32",
"assert",
"vecs",
".",
"ndim",
"==",
"2",
"N",
",",
"D",
"=",
"vecs",
".",
"shape",
"assert",
"D",
"==",
"self",
".",
"Ds",
"*",
"self",
".",
"M",
",",
"\"input dimension must be Ds * M\"",
"# codes[n][m] : code of n-th vec, m-th subspace",
"codes",
"=",
"np",
".",
"empty",
"(",
"(",
"N",
",",
"self",
".",
"M",
")",
",",
"dtype",
"=",
"self",
".",
"code_dtype",
")",
"for",
"m",
"in",
"range",
"(",
"self",
".",
"M",
")",
":",
"if",
"self",
".",
"verbose",
":",
"print",
"(",
"\"Encoding the subspace: {} / {}\"",
".",
"format",
"(",
"m",
",",
"self",
".",
"M",
")",
")",
"vecs_sub",
"=",
"vecs",
"[",
":",
",",
"m",
"*",
"self",
".",
"Ds",
":",
"(",
"m",
"+",
"1",
")",
"*",
"self",
".",
"Ds",
"]",
"codes",
"[",
":",
",",
"m",
"]",
",",
"_",
"=",
"vq",
"(",
"vecs_sub",
",",
"self",
".",
"codewords",
"[",
"m",
"]",
")",
"return",
"codes"
] | Encode input vectors into PQ-codes.
Args:
vecs (np.ndarray): Input vectors with shape=(N, D) and dtype=np.float32.
Returns:
np.ndarray: PQ codes with shape=(N, M) and dtype=self.code_dtype | [
"Encode",
"input",
"vectors",
"into",
"PQ",
"-",
"codes",
"."
] | 1ce68cad2e3cab62b409e6dd63f676ed7b443ee9 | https://github.com/matsui528/nanopq/blob/1ce68cad2e3cab62b409e6dd63f676ed7b443ee9/nanopq/pq.py#L89-L112 | train |
matsui528/nanopq | nanopq/pq.py | PQ.decode | def decode(self, codes):
"""Given PQ-codes, reconstruct original D-dimensional vectors
approximately by fetching the codewords.
Args:
codes (np.ndarray): PQ-cdoes with shape=(N, M) and dtype=self.code_dtype.
Each row is a PQ-code
Returns:
np.ndarray: Reconstructed vectors with shape=(N, D) and dtype=np.float32
"""
assert codes.ndim == 2
N, M = codes.shape
assert M == self.M
assert codes.dtype == self.code_dtype
vecs = np.empty((N, self.Ds * self.M), dtype=np.float32)
for m in range(self.M):
vecs[:, m * self.Ds : (m+1) * self.Ds] = self.codewords[m][codes[:, m], :]
return vecs | python | def decode(self, codes):
"""Given PQ-codes, reconstruct original D-dimensional vectors
approximately by fetching the codewords.
Args:
codes (np.ndarray): PQ-cdoes with shape=(N, M) and dtype=self.code_dtype.
Each row is a PQ-code
Returns:
np.ndarray: Reconstructed vectors with shape=(N, D) and dtype=np.float32
"""
assert codes.ndim == 2
N, M = codes.shape
assert M == self.M
assert codes.dtype == self.code_dtype
vecs = np.empty((N, self.Ds * self.M), dtype=np.float32)
for m in range(self.M):
vecs[:, m * self.Ds : (m+1) * self.Ds] = self.codewords[m][codes[:, m], :]
return vecs | [
"def",
"decode",
"(",
"self",
",",
"codes",
")",
":",
"assert",
"codes",
".",
"ndim",
"==",
"2",
"N",
",",
"M",
"=",
"codes",
".",
"shape",
"assert",
"M",
"==",
"self",
".",
"M",
"assert",
"codes",
".",
"dtype",
"==",
"self",
".",
"code_dtype",
"vecs",
"=",
"np",
".",
"empty",
"(",
"(",
"N",
",",
"self",
".",
"Ds",
"*",
"self",
".",
"M",
")",
",",
"dtype",
"=",
"np",
".",
"float32",
")",
"for",
"m",
"in",
"range",
"(",
"self",
".",
"M",
")",
":",
"vecs",
"[",
":",
",",
"m",
"*",
"self",
".",
"Ds",
":",
"(",
"m",
"+",
"1",
")",
"*",
"self",
".",
"Ds",
"]",
"=",
"self",
".",
"codewords",
"[",
"m",
"]",
"[",
"codes",
"[",
":",
",",
"m",
"]",
",",
":",
"]",
"return",
"vecs"
] | Given PQ-codes, reconstruct original D-dimensional vectors
approximately by fetching the codewords.
Args:
codes (np.ndarray): PQ-cdoes with shape=(N, M) and dtype=self.code_dtype.
Each row is a PQ-code
Returns:
np.ndarray: Reconstructed vectors with shape=(N, D) and dtype=np.float32 | [
"Given",
"PQ",
"-",
"codes",
"reconstruct",
"original",
"D",
"-",
"dimensional",
"vectors",
"approximately",
"by",
"fetching",
"the",
"codewords",
"."
] | 1ce68cad2e3cab62b409e6dd63f676ed7b443ee9 | https://github.com/matsui528/nanopq/blob/1ce68cad2e3cab62b409e6dd63f676ed7b443ee9/nanopq/pq.py#L114-L135 | train |
stephenmcd/hot-redis | hot_redis/client.py | transaction | def transaction():
"""
Swaps out the current client with a pipeline instance,
so that each Redis method call inside the context will be
pipelined. Once the context is exited, we execute the pipeline.
"""
client = default_client()
_thread.client = client.pipeline()
try:
yield
_thread.client.execute()
finally:
_thread.client = client | python | def transaction():
"""
Swaps out the current client with a pipeline instance,
so that each Redis method call inside the context will be
pipelined. Once the context is exited, we execute the pipeline.
"""
client = default_client()
_thread.client = client.pipeline()
try:
yield
_thread.client.execute()
finally:
_thread.client = client | [
"def",
"transaction",
"(",
")",
":",
"client",
"=",
"default_client",
"(",
")",
"_thread",
".",
"client",
"=",
"client",
".",
"pipeline",
"(",
")",
"try",
":",
"yield",
"_thread",
".",
"client",
".",
"execute",
"(",
")",
"finally",
":",
"_thread",
".",
"client",
"=",
"client"
] | Swaps out the current client with a pipeline instance,
so that each Redis method call inside the context will be
pipelined. Once the context is exited, we execute the pipeline. | [
"Swaps",
"out",
"the",
"current",
"client",
"with",
"a",
"pipeline",
"instance",
"so",
"that",
"each",
"Redis",
"method",
"call",
"inside",
"the",
"context",
"will",
"be",
"pipelined",
".",
"Once",
"the",
"context",
"is",
"exited",
"we",
"execute",
"the",
"pipeline",
"."
] | 6b0cf260c775fd98c44b6703030d33004dabf67d | https://github.com/stephenmcd/hot-redis/blob/6b0cf260c775fd98c44b6703030d33004dabf67d/hot_redis/client.py#L78-L90 | train |
stephenmcd/hot-redis | hot_redis/client.py | HotClient._get_lua_path | def _get_lua_path(self, name):
"""
Joins the given name with the relative path of the module.
"""
parts = (os.path.dirname(os.path.abspath(__file__)), "lua", name)
return os.path.join(*parts) | python | def _get_lua_path(self, name):
"""
Joins the given name with the relative path of the module.
"""
parts = (os.path.dirname(os.path.abspath(__file__)), "lua", name)
return os.path.join(*parts) | [
"def",
"_get_lua_path",
"(",
"self",
",",
"name",
")",
":",
"parts",
"=",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"os",
".",
"path",
".",
"abspath",
"(",
"__file__",
")",
")",
",",
"\"lua\"",
",",
"name",
")",
"return",
"os",
".",
"path",
".",
"join",
"(",
"*",
"parts",
")"
] | Joins the given name with the relative path of the module. | [
"Joins",
"the",
"given",
"name",
"with",
"the",
"relative",
"path",
"of",
"the",
"module",
"."
] | 6b0cf260c775fd98c44b6703030d33004dabf67d | https://github.com/stephenmcd/hot-redis/blob/6b0cf260c775fd98c44b6703030d33004dabf67d/hot_redis/client.py#L27-L32 | train |
stephenmcd/hot-redis | hot_redis/client.py | HotClient._create_lua_method | def _create_lua_method(self, name, code):
"""
Registers the code snippet as a Lua script, and binds the
script to the client as a method that can be called with
the same signature as regular client methods, eg with a
single key arg.
"""
script = self.register_script(code)
setattr(script, "name", name) # Helps debugging redis lib.
method = lambda key, *a, **k: script(keys=[key], args=a, **k)
setattr(self, name, method) | python | def _create_lua_method(self, name, code):
"""
Registers the code snippet as a Lua script, and binds the
script to the client as a method that can be called with
the same signature as regular client methods, eg with a
single key arg.
"""
script = self.register_script(code)
setattr(script, "name", name) # Helps debugging redis lib.
method = lambda key, *a, **k: script(keys=[key], args=a, **k)
setattr(self, name, method) | [
"def",
"_create_lua_method",
"(",
"self",
",",
"name",
",",
"code",
")",
":",
"script",
"=",
"self",
".",
"register_script",
"(",
"code",
")",
"setattr",
"(",
"script",
",",
"\"name\"",
",",
"name",
")",
"# Helps debugging redis lib.",
"method",
"=",
"lambda",
"key",
",",
"*",
"a",
",",
"*",
"*",
"k",
":",
"script",
"(",
"keys",
"=",
"[",
"key",
"]",
",",
"args",
"=",
"a",
",",
"*",
"*",
"k",
")",
"setattr",
"(",
"self",
",",
"name",
",",
"method",
")"
] | Registers the code snippet as a Lua script, and binds the
script to the client as a method that can be called with
the same signature as regular client methods, eg with a
single key arg. | [
"Registers",
"the",
"code",
"snippet",
"as",
"a",
"Lua",
"script",
"and",
"binds",
"the",
"script",
"to",
"the",
"client",
"as",
"a",
"method",
"that",
"can",
"be",
"called",
"with",
"the",
"same",
"signature",
"as",
"regular",
"client",
"methods",
"eg",
"with",
"a",
"single",
"key",
"arg",
"."
] | 6b0cf260c775fd98c44b6703030d33004dabf67d | https://github.com/stephenmcd/hot-redis/blob/6b0cf260c775fd98c44b6703030d33004dabf67d/hot_redis/client.py#L47-L57 | train |
stephenmcd/hot-redis | hot_redis/types.py | value_left | def value_left(self, other):
"""
Returns the value of the other type instance to use in an
operator method, namely when the method's instance is on the
left side of the expression.
"""
return other.value if isinstance(other, self.__class__) else other | python | def value_left(self, other):
"""
Returns the value of the other type instance to use in an
operator method, namely when the method's instance is on the
left side of the expression.
"""
return other.value if isinstance(other, self.__class__) else other | [
"def",
"value_left",
"(",
"self",
",",
"other",
")",
":",
"return",
"other",
".",
"value",
"if",
"isinstance",
"(",
"other",
",",
"self",
".",
"__class__",
")",
"else",
"other"
] | Returns the value of the other type instance to use in an
operator method, namely when the method's instance is on the
left side of the expression. | [
"Returns",
"the",
"value",
"of",
"the",
"other",
"type",
"instance",
"to",
"use",
"in",
"an",
"operator",
"method",
"namely",
"when",
"the",
"method",
"s",
"instance",
"is",
"on",
"the",
"left",
"side",
"of",
"the",
"expression",
"."
] | 6b0cf260c775fd98c44b6703030d33004dabf67d | https://github.com/stephenmcd/hot-redis/blob/6b0cf260c775fd98c44b6703030d33004dabf67d/hot_redis/types.py#L28-L34 | train |
stephenmcd/hot-redis | hot_redis/types.py | value_right | def value_right(self, other):
"""
Returns the value of the type instance calling an to use in an
operator method, namely when the method's instance is on the
right side of the expression.
"""
return self if isinstance(other, self.__class__) else self.value | python | def value_right(self, other):
"""
Returns the value of the type instance calling an to use in an
operator method, namely when the method's instance is on the
right side of the expression.
"""
return self if isinstance(other, self.__class__) else self.value | [
"def",
"value_right",
"(",
"self",
",",
"other",
")",
":",
"return",
"self",
"if",
"isinstance",
"(",
"other",
",",
"self",
".",
"__class__",
")",
"else",
"self",
".",
"value"
] | Returns the value of the type instance calling an to use in an
operator method, namely when the method's instance is on the
right side of the expression. | [
"Returns",
"the",
"value",
"of",
"the",
"type",
"instance",
"calling",
"an",
"to",
"use",
"in",
"an",
"operator",
"method",
"namely",
"when",
"the",
"method",
"s",
"instance",
"is",
"on",
"the",
"right",
"side",
"of",
"the",
"expression",
"."
] | 6b0cf260c775fd98c44b6703030d33004dabf67d | https://github.com/stephenmcd/hot-redis/blob/6b0cf260c775fd98c44b6703030d33004dabf67d/hot_redis/types.py#L37-L43 | train |
stephenmcd/hot-redis | hot_redis/types.py | op_left | def op_left(op):
"""
Returns a type instance method for the given operator, applied
when the instance appears on the left side of the expression.
"""
def method(self, other):
return op(self.value, value_left(self, other))
return method | python | def op_left(op):
"""
Returns a type instance method for the given operator, applied
when the instance appears on the left side of the expression.
"""
def method(self, other):
return op(self.value, value_left(self, other))
return method | [
"def",
"op_left",
"(",
"op",
")",
":",
"def",
"method",
"(",
"self",
",",
"other",
")",
":",
"return",
"op",
"(",
"self",
".",
"value",
",",
"value_left",
"(",
"self",
",",
"other",
")",
")",
"return",
"method"
] | Returns a type instance method for the given operator, applied
when the instance appears on the left side of the expression. | [
"Returns",
"a",
"type",
"instance",
"method",
"for",
"the",
"given",
"operator",
"applied",
"when",
"the",
"instance",
"appears",
"on",
"the",
"left",
"side",
"of",
"the",
"expression",
"."
] | 6b0cf260c775fd98c44b6703030d33004dabf67d | https://github.com/stephenmcd/hot-redis/blob/6b0cf260c775fd98c44b6703030d33004dabf67d/hot_redis/types.py#L46-L53 | train |
stephenmcd/hot-redis | hot_redis/types.py | op_right | def op_right(op):
"""
Returns a type instance method for the given operator, applied
when the instance appears on the right side of the expression.
"""
def method(self, other):
return op(value_left(self, other), value_right(self, other))
return method | python | def op_right(op):
"""
Returns a type instance method for the given operator, applied
when the instance appears on the right side of the expression.
"""
def method(self, other):
return op(value_left(self, other), value_right(self, other))
return method | [
"def",
"op_right",
"(",
"op",
")",
":",
"def",
"method",
"(",
"self",
",",
"other",
")",
":",
"return",
"op",
"(",
"value_left",
"(",
"self",
",",
"other",
")",
",",
"value_right",
"(",
"self",
",",
"other",
")",
")",
"return",
"method"
] | Returns a type instance method for the given operator, applied
when the instance appears on the right side of the expression. | [
"Returns",
"a",
"type",
"instance",
"method",
"for",
"the",
"given",
"operator",
"applied",
"when",
"the",
"instance",
"appears",
"on",
"the",
"right",
"side",
"of",
"the",
"expression",
"."
] | 6b0cf260c775fd98c44b6703030d33004dabf67d | https://github.com/stephenmcd/hot-redis/blob/6b0cf260c775fd98c44b6703030d33004dabf67d/hot_redis/types.py#L56-L63 | train |
jfhbrook/pyee | pyee/_base.py | BaseEventEmitter.on | def on(self, event, f=None):
"""Registers the function ``f`` to the event name ``event``.
If ``f`` isn't provided, this method returns a function that
takes ``f`` as a callback; in other words, you can use this method
as a decorator, like so::
@ee.on('data')
def data_handler(data):
print(data)
In both the decorated and undecorated forms, the event handler is
returned. The upshot of this is that you can call decorated handlers
directly, as well as use them in remove_listener calls.
"""
def _on(f):
self._add_event_handler(event, f, f)
return f
if f is None:
return _on
else:
return _on(f) | python | def on(self, event, f=None):
"""Registers the function ``f`` to the event name ``event``.
If ``f`` isn't provided, this method returns a function that
takes ``f`` as a callback; in other words, you can use this method
as a decorator, like so::
@ee.on('data')
def data_handler(data):
print(data)
In both the decorated and undecorated forms, the event handler is
returned. The upshot of this is that you can call decorated handlers
directly, as well as use them in remove_listener calls.
"""
def _on(f):
self._add_event_handler(event, f, f)
return f
if f is None:
return _on
else:
return _on(f) | [
"def",
"on",
"(",
"self",
",",
"event",
",",
"f",
"=",
"None",
")",
":",
"def",
"_on",
"(",
"f",
")",
":",
"self",
".",
"_add_event_handler",
"(",
"event",
",",
"f",
",",
"f",
")",
"return",
"f",
"if",
"f",
"is",
"None",
":",
"return",
"_on",
"else",
":",
"return",
"_on",
"(",
"f",
")"
] | Registers the function ``f`` to the event name ``event``.
If ``f`` isn't provided, this method returns a function that
takes ``f`` as a callback; in other words, you can use this method
as a decorator, like so::
@ee.on('data')
def data_handler(data):
print(data)
In both the decorated and undecorated forms, the event handler is
returned. The upshot of this is that you can call decorated handlers
directly, as well as use them in remove_listener calls. | [
"Registers",
"the",
"function",
"f",
"to",
"the",
"event",
"name",
"event",
"."
] | f5896239f421381e8b787c61c6185e33defb0292 | https://github.com/jfhbrook/pyee/blob/f5896239f421381e8b787c61c6185e33defb0292/pyee/_base.py#L42-L65 | train |
jfhbrook/pyee | pyee/_base.py | BaseEventEmitter.once | def once(self, event, f=None):
"""The same as ``ee.on``, except that the listener is automatically
removed after being called.
"""
def _wrapper(f):
def g(*args, **kwargs):
self.remove_listener(event, f)
# f may return a coroutine, so we need to return that
# result here so that emit can schedule it
return f(*args, **kwargs)
self._add_event_handler(event, f, g)
return f
if f is None:
return _wrapper
else:
return _wrapper(f) | python | def once(self, event, f=None):
"""The same as ``ee.on``, except that the listener is automatically
removed after being called.
"""
def _wrapper(f):
def g(*args, **kwargs):
self.remove_listener(event, f)
# f may return a coroutine, so we need to return that
# result here so that emit can schedule it
return f(*args, **kwargs)
self._add_event_handler(event, f, g)
return f
if f is None:
return _wrapper
else:
return _wrapper(f) | [
"def",
"once",
"(",
"self",
",",
"event",
",",
"f",
"=",
"None",
")",
":",
"def",
"_wrapper",
"(",
"f",
")",
":",
"def",
"g",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"self",
".",
"remove_listener",
"(",
"event",
",",
"f",
")",
"# f may return a coroutine, so we need to return that",
"# result here so that emit can schedule it",
"return",
"f",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"self",
".",
"_add_event_handler",
"(",
"event",
",",
"f",
",",
"g",
")",
"return",
"f",
"if",
"f",
"is",
"None",
":",
"return",
"_wrapper",
"else",
":",
"return",
"_wrapper",
"(",
"f",
")"
] | The same as ``ee.on``, except that the listener is automatically
removed after being called. | [
"The",
"same",
"as",
"ee",
".",
"on",
"except",
"that",
"the",
"listener",
"is",
"automatically",
"removed",
"after",
"being",
"called",
"."
] | f5896239f421381e8b787c61c6185e33defb0292 | https://github.com/jfhbrook/pyee/blob/f5896239f421381e8b787c61c6185e33defb0292/pyee/_base.py#L110-L127 | train |
jfhbrook/pyee | pyee/_base.py | BaseEventEmitter.remove_all_listeners | def remove_all_listeners(self, event=None):
"""Remove all listeners attached to ``event``.
If ``event`` is ``None``, remove all listeners on all events.
"""
if event is not None:
self._events[event] = OrderedDict()
else:
self._events = defaultdict(OrderedDict) | python | def remove_all_listeners(self, event=None):
"""Remove all listeners attached to ``event``.
If ``event`` is ``None``, remove all listeners on all events.
"""
if event is not None:
self._events[event] = OrderedDict()
else:
self._events = defaultdict(OrderedDict) | [
"def",
"remove_all_listeners",
"(",
"self",
",",
"event",
"=",
"None",
")",
":",
"if",
"event",
"is",
"not",
"None",
":",
"self",
".",
"_events",
"[",
"event",
"]",
"=",
"OrderedDict",
"(",
")",
"else",
":",
"self",
".",
"_events",
"=",
"defaultdict",
"(",
"OrderedDict",
")"
] | Remove all listeners attached to ``event``.
If ``event`` is ``None``, remove all listeners on all events. | [
"Remove",
"all",
"listeners",
"attached",
"to",
"event",
".",
"If",
"event",
"is",
"None",
"remove",
"all",
"listeners",
"on",
"all",
"events",
"."
] | f5896239f421381e8b787c61c6185e33defb0292 | https://github.com/jfhbrook/pyee/blob/f5896239f421381e8b787c61c6185e33defb0292/pyee/_base.py#L133-L140 | train |
scott-griffiths/bitstring | bitstring.py | offsetcopy | def offsetcopy(s, newoffset):
"""Return a copy of a ByteStore with the newoffset.
Not part of public interface.
"""
assert 0 <= newoffset < 8
if not s.bitlength:
return copy.copy(s)
else:
if newoffset == s.offset % 8:
return ByteStore(s.getbyteslice(s.byteoffset, s.byteoffset + s.bytelength), s.bitlength, newoffset)
newdata = []
d = s._rawarray
assert newoffset != s.offset % 8
if newoffset < s.offset % 8:
# We need to shift everything left
shiftleft = s.offset % 8 - newoffset
# First deal with everything except for the final byte
for x in range(s.byteoffset, s.byteoffset + s.bytelength - 1):
newdata.append(((d[x] << shiftleft) & 0xff) +\
(d[x + 1] >> (8 - shiftleft)))
bits_in_last_byte = (s.offset + s.bitlength) % 8
if not bits_in_last_byte:
bits_in_last_byte = 8
if bits_in_last_byte > shiftleft:
newdata.append((d[s.byteoffset + s.bytelength - 1] << shiftleft) & 0xff)
else: # newoffset > s._offset % 8
shiftright = newoffset - s.offset % 8
newdata.append(s.getbyte(0) >> shiftright)
for x in range(s.byteoffset + 1, s.byteoffset + s.bytelength):
newdata.append(((d[x - 1] << (8 - shiftright)) & 0xff) +\
(d[x] >> shiftright))
bits_in_last_byte = (s.offset + s.bitlength) % 8
if not bits_in_last_byte:
bits_in_last_byte = 8
if bits_in_last_byte + shiftright > 8:
newdata.append((d[s.byteoffset + s.bytelength - 1] << (8 - shiftright)) & 0xff)
new_s = ByteStore(bytearray(newdata), s.bitlength, newoffset)
assert new_s.offset == newoffset
return new_s | python | def offsetcopy(s, newoffset):
"""Return a copy of a ByteStore with the newoffset.
Not part of public interface.
"""
assert 0 <= newoffset < 8
if not s.bitlength:
return copy.copy(s)
else:
if newoffset == s.offset % 8:
return ByteStore(s.getbyteslice(s.byteoffset, s.byteoffset + s.bytelength), s.bitlength, newoffset)
newdata = []
d = s._rawarray
assert newoffset != s.offset % 8
if newoffset < s.offset % 8:
# We need to shift everything left
shiftleft = s.offset % 8 - newoffset
# First deal with everything except for the final byte
for x in range(s.byteoffset, s.byteoffset + s.bytelength - 1):
newdata.append(((d[x] << shiftleft) & 0xff) +\
(d[x + 1] >> (8 - shiftleft)))
bits_in_last_byte = (s.offset + s.bitlength) % 8
if not bits_in_last_byte:
bits_in_last_byte = 8
if bits_in_last_byte > shiftleft:
newdata.append((d[s.byteoffset + s.bytelength - 1] << shiftleft) & 0xff)
else: # newoffset > s._offset % 8
shiftright = newoffset - s.offset % 8
newdata.append(s.getbyte(0) >> shiftright)
for x in range(s.byteoffset + 1, s.byteoffset + s.bytelength):
newdata.append(((d[x - 1] << (8 - shiftright)) & 0xff) +\
(d[x] >> shiftright))
bits_in_last_byte = (s.offset + s.bitlength) % 8
if not bits_in_last_byte:
bits_in_last_byte = 8
if bits_in_last_byte + shiftright > 8:
newdata.append((d[s.byteoffset + s.bytelength - 1] << (8 - shiftright)) & 0xff)
new_s = ByteStore(bytearray(newdata), s.bitlength, newoffset)
assert new_s.offset == newoffset
return new_s | [
"def",
"offsetcopy",
"(",
"s",
",",
"newoffset",
")",
":",
"assert",
"0",
"<=",
"newoffset",
"<",
"8",
"if",
"not",
"s",
".",
"bitlength",
":",
"return",
"copy",
".",
"copy",
"(",
"s",
")",
"else",
":",
"if",
"newoffset",
"==",
"s",
".",
"offset",
"%",
"8",
":",
"return",
"ByteStore",
"(",
"s",
".",
"getbyteslice",
"(",
"s",
".",
"byteoffset",
",",
"s",
".",
"byteoffset",
"+",
"s",
".",
"bytelength",
")",
",",
"s",
".",
"bitlength",
",",
"newoffset",
")",
"newdata",
"=",
"[",
"]",
"d",
"=",
"s",
".",
"_rawarray",
"assert",
"newoffset",
"!=",
"s",
".",
"offset",
"%",
"8",
"if",
"newoffset",
"<",
"s",
".",
"offset",
"%",
"8",
":",
"# We need to shift everything left",
"shiftleft",
"=",
"s",
".",
"offset",
"%",
"8",
"-",
"newoffset",
"# First deal with everything except for the final byte",
"for",
"x",
"in",
"range",
"(",
"s",
".",
"byteoffset",
",",
"s",
".",
"byteoffset",
"+",
"s",
".",
"bytelength",
"-",
"1",
")",
":",
"newdata",
".",
"append",
"(",
"(",
"(",
"d",
"[",
"x",
"]",
"<<",
"shiftleft",
")",
"&",
"0xff",
")",
"+",
"(",
"d",
"[",
"x",
"+",
"1",
"]",
">>",
"(",
"8",
"-",
"shiftleft",
")",
")",
")",
"bits_in_last_byte",
"=",
"(",
"s",
".",
"offset",
"+",
"s",
".",
"bitlength",
")",
"%",
"8",
"if",
"not",
"bits_in_last_byte",
":",
"bits_in_last_byte",
"=",
"8",
"if",
"bits_in_last_byte",
">",
"shiftleft",
":",
"newdata",
".",
"append",
"(",
"(",
"d",
"[",
"s",
".",
"byteoffset",
"+",
"s",
".",
"bytelength",
"-",
"1",
"]",
"<<",
"shiftleft",
")",
"&",
"0xff",
")",
"else",
":",
"# newoffset > s._offset % 8",
"shiftright",
"=",
"newoffset",
"-",
"s",
".",
"offset",
"%",
"8",
"newdata",
".",
"append",
"(",
"s",
".",
"getbyte",
"(",
"0",
")",
">>",
"shiftright",
")",
"for",
"x",
"in",
"range",
"(",
"s",
".",
"byteoffset",
"+",
"1",
",",
"s",
".",
"byteoffset",
"+",
"s",
".",
"bytelength",
")",
":",
"newdata",
".",
"append",
"(",
"(",
"(",
"d",
"[",
"x",
"-",
"1",
"]",
"<<",
"(",
"8",
"-",
"shiftright",
")",
")",
"&",
"0xff",
")",
"+",
"(",
"d",
"[",
"x",
"]",
">>",
"shiftright",
")",
")",
"bits_in_last_byte",
"=",
"(",
"s",
".",
"offset",
"+",
"s",
".",
"bitlength",
")",
"%",
"8",
"if",
"not",
"bits_in_last_byte",
":",
"bits_in_last_byte",
"=",
"8",
"if",
"bits_in_last_byte",
"+",
"shiftright",
">",
"8",
":",
"newdata",
".",
"append",
"(",
"(",
"d",
"[",
"s",
".",
"byteoffset",
"+",
"s",
".",
"bytelength",
"-",
"1",
"]",
"<<",
"(",
"8",
"-",
"shiftright",
")",
")",
"&",
"0xff",
")",
"new_s",
"=",
"ByteStore",
"(",
"bytearray",
"(",
"newdata",
")",
",",
"s",
".",
"bitlength",
",",
"newoffset",
")",
"assert",
"new_s",
".",
"offset",
"==",
"newoffset",
"return",
"new_s"
] | Return a copy of a ByteStore with the newoffset.
Not part of public interface. | [
"Return",
"a",
"copy",
"of",
"a",
"ByteStore",
"with",
"the",
"newoffset",
"."
] | ab40ae7f0b43fe223a39b63cbc0529b09f3ef653 | https://github.com/scott-griffiths/bitstring/blob/ab40ae7f0b43fe223a39b63cbc0529b09f3ef653/bitstring.py#L248-L287 | train |
scott-griffiths/bitstring | bitstring.py | structparser | def structparser(token):
"""Parse struct-like format string token into sub-token list."""
m = STRUCT_PACK_RE.match(token)
if not m:
return [token]
else:
endian = m.group('endian')
if endian is None:
return [token]
# Split the format string into a list of 'q', '4h' etc.
formatlist = re.findall(STRUCT_SPLIT_RE, m.group('fmt'))
# Now deal with mulitiplicative factors, 4h -> hhhh etc.
fmt = ''.join([f[-1] * int(f[:-1]) if len(f) != 1 else
f for f in formatlist])
if endian == '@':
# Native endianness
if byteorder == 'little':
endian = '<'
else:
assert byteorder == 'big'
endian = '>'
if endian == '<':
tokens = [REPLACEMENTS_LE[c] for c in fmt]
else:
assert endian == '>'
tokens = [REPLACEMENTS_BE[c] for c in fmt]
return tokens | python | def structparser(token):
"""Parse struct-like format string token into sub-token list."""
m = STRUCT_PACK_RE.match(token)
if not m:
return [token]
else:
endian = m.group('endian')
if endian is None:
return [token]
# Split the format string into a list of 'q', '4h' etc.
formatlist = re.findall(STRUCT_SPLIT_RE, m.group('fmt'))
# Now deal with mulitiplicative factors, 4h -> hhhh etc.
fmt = ''.join([f[-1] * int(f[:-1]) if len(f) != 1 else
f for f in formatlist])
if endian == '@':
# Native endianness
if byteorder == 'little':
endian = '<'
else:
assert byteorder == 'big'
endian = '>'
if endian == '<':
tokens = [REPLACEMENTS_LE[c] for c in fmt]
else:
assert endian == '>'
tokens = [REPLACEMENTS_BE[c] for c in fmt]
return tokens | [
"def",
"structparser",
"(",
"token",
")",
":",
"m",
"=",
"STRUCT_PACK_RE",
".",
"match",
"(",
"token",
")",
"if",
"not",
"m",
":",
"return",
"[",
"token",
"]",
"else",
":",
"endian",
"=",
"m",
".",
"group",
"(",
"'endian'",
")",
"if",
"endian",
"is",
"None",
":",
"return",
"[",
"token",
"]",
"# Split the format string into a list of 'q', '4h' etc.",
"formatlist",
"=",
"re",
".",
"findall",
"(",
"STRUCT_SPLIT_RE",
",",
"m",
".",
"group",
"(",
"'fmt'",
")",
")",
"# Now deal with mulitiplicative factors, 4h -> hhhh etc.",
"fmt",
"=",
"''",
".",
"join",
"(",
"[",
"f",
"[",
"-",
"1",
"]",
"*",
"int",
"(",
"f",
"[",
":",
"-",
"1",
"]",
")",
"if",
"len",
"(",
"f",
")",
"!=",
"1",
"else",
"f",
"for",
"f",
"in",
"formatlist",
"]",
")",
"if",
"endian",
"==",
"'@'",
":",
"# Native endianness",
"if",
"byteorder",
"==",
"'little'",
":",
"endian",
"=",
"'<'",
"else",
":",
"assert",
"byteorder",
"==",
"'big'",
"endian",
"=",
"'>'",
"if",
"endian",
"==",
"'<'",
":",
"tokens",
"=",
"[",
"REPLACEMENTS_LE",
"[",
"c",
"]",
"for",
"c",
"in",
"fmt",
"]",
"else",
":",
"assert",
"endian",
"==",
"'>'",
"tokens",
"=",
"[",
"REPLACEMENTS_BE",
"[",
"c",
"]",
"for",
"c",
"in",
"fmt",
"]",
"return",
"tokens"
] | Parse struct-like format string token into sub-token list. | [
"Parse",
"struct",
"-",
"like",
"format",
"string",
"token",
"into",
"sub",
"-",
"token",
"list",
"."
] | ab40ae7f0b43fe223a39b63cbc0529b09f3ef653 | https://github.com/scott-griffiths/bitstring/blob/ab40ae7f0b43fe223a39b63cbc0529b09f3ef653/bitstring.py#L506-L532 | train |
scott-griffiths/bitstring | bitstring.py | tokenparser | def tokenparser(fmt, keys=None, token_cache={}):
"""Divide the format string into tokens and parse them.
Return stretchy token and list of [initialiser, length, value]
initialiser is one of: hex, oct, bin, uint, int, se, ue, 0x, 0o, 0b etc.
length is None if not known, as is value.
If the token is in the keyword dictionary (keys) then it counts as a
special case and isn't messed with.
tokens must be of the form: [factor*][initialiser][:][length][=value]
"""
try:
return token_cache[(fmt, keys)]
except KeyError:
token_key = (fmt, keys)
# Very inefficient expanding of brackets.
fmt = expand_brackets(fmt)
# Split tokens by ',' and remove whitespace
# The meta_tokens can either be ordinary single tokens or multiple
# struct-format token strings.
meta_tokens = (''.join(f.split()) for f in fmt.split(','))
return_values = []
stretchy_token = False
for meta_token in meta_tokens:
# See if it has a multiplicative factor
m = MULTIPLICATIVE_RE.match(meta_token)
if not m:
factor = 1
else:
factor = int(m.group('factor'))
meta_token = m.group('token')
# See if it's a struct-like format
tokens = structparser(meta_token)
ret_vals = []
for token in tokens:
if keys and token in keys:
# Don't bother parsing it, it's a keyword argument
ret_vals.append([token, None, None])
continue
value = length = None
if token == '':
continue
# Match literal tokens of the form 0x... 0o... and 0b...
m = LITERAL_RE.match(token)
if m:
name = m.group('name')
value = m.group('value')
ret_vals.append([name, length, value])
continue
# Match everything else:
m1 = TOKEN_RE.match(token)
if not m1:
# and if you don't specify a 'name' then the default is 'uint':
m2 = DEFAULT_UINT.match(token)
if not m2:
raise ValueError("Don't understand token '{0}'.".format(token))
if m1:
name = m1.group('name')
length = m1.group('len')
if m1.group('value'):
value = m1.group('value')
else:
assert m2
name = 'uint'
length = m2.group('len')
if m2.group('value'):
value = m2.group('value')
if name == 'bool':
if length is not None:
raise ValueError("You can't specify a length with bool tokens - they are always one bit.")
length = 1
if length is None and name not in ('se', 'ue', 'sie', 'uie'):
stretchy_token = True
if length is not None:
# Try converting length to int, otherwise check it's a key.
try:
length = int(length)
if length < 0:
raise Error
# For the 'bytes' token convert length to bits.
if name == 'bytes':
length *= 8
except Error:
raise ValueError("Can't read a token with a negative length.")
except ValueError:
if not keys or length not in keys:
raise ValueError("Don't understand length '{0}' of token.".format(length))
ret_vals.append([name, length, value])
# This multiplies by the multiplicative factor, but this means that
# we can't allow keyword values as multipliers (e.g. n*uint:8).
# The only way to do this would be to return the factor in some fashion
# (we can't use the key's value here as it would mean that we couldn't
# sensibly continue to cache the function's results. (TODO).
return_values.extend(ret_vals * factor)
return_values = [tuple(x) for x in return_values]
if len(token_cache) < CACHE_SIZE:
token_cache[token_key] = stretchy_token, return_values
return stretchy_token, return_values | python | def tokenparser(fmt, keys=None, token_cache={}):
"""Divide the format string into tokens and parse them.
Return stretchy token and list of [initialiser, length, value]
initialiser is one of: hex, oct, bin, uint, int, se, ue, 0x, 0o, 0b etc.
length is None if not known, as is value.
If the token is in the keyword dictionary (keys) then it counts as a
special case and isn't messed with.
tokens must be of the form: [factor*][initialiser][:][length][=value]
"""
try:
return token_cache[(fmt, keys)]
except KeyError:
token_key = (fmt, keys)
# Very inefficient expanding of brackets.
fmt = expand_brackets(fmt)
# Split tokens by ',' and remove whitespace
# The meta_tokens can either be ordinary single tokens or multiple
# struct-format token strings.
meta_tokens = (''.join(f.split()) for f in fmt.split(','))
return_values = []
stretchy_token = False
for meta_token in meta_tokens:
# See if it has a multiplicative factor
m = MULTIPLICATIVE_RE.match(meta_token)
if not m:
factor = 1
else:
factor = int(m.group('factor'))
meta_token = m.group('token')
# See if it's a struct-like format
tokens = structparser(meta_token)
ret_vals = []
for token in tokens:
if keys and token in keys:
# Don't bother parsing it, it's a keyword argument
ret_vals.append([token, None, None])
continue
value = length = None
if token == '':
continue
# Match literal tokens of the form 0x... 0o... and 0b...
m = LITERAL_RE.match(token)
if m:
name = m.group('name')
value = m.group('value')
ret_vals.append([name, length, value])
continue
# Match everything else:
m1 = TOKEN_RE.match(token)
if not m1:
# and if you don't specify a 'name' then the default is 'uint':
m2 = DEFAULT_UINT.match(token)
if not m2:
raise ValueError("Don't understand token '{0}'.".format(token))
if m1:
name = m1.group('name')
length = m1.group('len')
if m1.group('value'):
value = m1.group('value')
else:
assert m2
name = 'uint'
length = m2.group('len')
if m2.group('value'):
value = m2.group('value')
if name == 'bool':
if length is not None:
raise ValueError("You can't specify a length with bool tokens - they are always one bit.")
length = 1
if length is None and name not in ('se', 'ue', 'sie', 'uie'):
stretchy_token = True
if length is not None:
# Try converting length to int, otherwise check it's a key.
try:
length = int(length)
if length < 0:
raise Error
# For the 'bytes' token convert length to bits.
if name == 'bytes':
length *= 8
except Error:
raise ValueError("Can't read a token with a negative length.")
except ValueError:
if not keys or length not in keys:
raise ValueError("Don't understand length '{0}' of token.".format(length))
ret_vals.append([name, length, value])
# This multiplies by the multiplicative factor, but this means that
# we can't allow keyword values as multipliers (e.g. n*uint:8).
# The only way to do this would be to return the factor in some fashion
# (we can't use the key's value here as it would mean that we couldn't
# sensibly continue to cache the function's results. (TODO).
return_values.extend(ret_vals * factor)
return_values = [tuple(x) for x in return_values]
if len(token_cache) < CACHE_SIZE:
token_cache[token_key] = stretchy_token, return_values
return stretchy_token, return_values | [
"def",
"tokenparser",
"(",
"fmt",
",",
"keys",
"=",
"None",
",",
"token_cache",
"=",
"{",
"}",
")",
":",
"try",
":",
"return",
"token_cache",
"[",
"(",
"fmt",
",",
"keys",
")",
"]",
"except",
"KeyError",
":",
"token_key",
"=",
"(",
"fmt",
",",
"keys",
")",
"# Very inefficient expanding of brackets.",
"fmt",
"=",
"expand_brackets",
"(",
"fmt",
")",
"# Split tokens by ',' and remove whitespace",
"# The meta_tokens can either be ordinary single tokens or multiple",
"# struct-format token strings.",
"meta_tokens",
"=",
"(",
"''",
".",
"join",
"(",
"f",
".",
"split",
"(",
")",
")",
"for",
"f",
"in",
"fmt",
".",
"split",
"(",
"','",
")",
")",
"return_values",
"=",
"[",
"]",
"stretchy_token",
"=",
"False",
"for",
"meta_token",
"in",
"meta_tokens",
":",
"# See if it has a multiplicative factor",
"m",
"=",
"MULTIPLICATIVE_RE",
".",
"match",
"(",
"meta_token",
")",
"if",
"not",
"m",
":",
"factor",
"=",
"1",
"else",
":",
"factor",
"=",
"int",
"(",
"m",
".",
"group",
"(",
"'factor'",
")",
")",
"meta_token",
"=",
"m",
".",
"group",
"(",
"'token'",
")",
"# See if it's a struct-like format",
"tokens",
"=",
"structparser",
"(",
"meta_token",
")",
"ret_vals",
"=",
"[",
"]",
"for",
"token",
"in",
"tokens",
":",
"if",
"keys",
"and",
"token",
"in",
"keys",
":",
"# Don't bother parsing it, it's a keyword argument",
"ret_vals",
".",
"append",
"(",
"[",
"token",
",",
"None",
",",
"None",
"]",
")",
"continue",
"value",
"=",
"length",
"=",
"None",
"if",
"token",
"==",
"''",
":",
"continue",
"# Match literal tokens of the form 0x... 0o... and 0b...",
"m",
"=",
"LITERAL_RE",
".",
"match",
"(",
"token",
")",
"if",
"m",
":",
"name",
"=",
"m",
".",
"group",
"(",
"'name'",
")",
"value",
"=",
"m",
".",
"group",
"(",
"'value'",
")",
"ret_vals",
".",
"append",
"(",
"[",
"name",
",",
"length",
",",
"value",
"]",
")",
"continue",
"# Match everything else:",
"m1",
"=",
"TOKEN_RE",
".",
"match",
"(",
"token",
")",
"if",
"not",
"m1",
":",
"# and if you don't specify a 'name' then the default is 'uint':",
"m2",
"=",
"DEFAULT_UINT",
".",
"match",
"(",
"token",
")",
"if",
"not",
"m2",
":",
"raise",
"ValueError",
"(",
"\"Don't understand token '{0}'.\"",
".",
"format",
"(",
"token",
")",
")",
"if",
"m1",
":",
"name",
"=",
"m1",
".",
"group",
"(",
"'name'",
")",
"length",
"=",
"m1",
".",
"group",
"(",
"'len'",
")",
"if",
"m1",
".",
"group",
"(",
"'value'",
")",
":",
"value",
"=",
"m1",
".",
"group",
"(",
"'value'",
")",
"else",
":",
"assert",
"m2",
"name",
"=",
"'uint'",
"length",
"=",
"m2",
".",
"group",
"(",
"'len'",
")",
"if",
"m2",
".",
"group",
"(",
"'value'",
")",
":",
"value",
"=",
"m2",
".",
"group",
"(",
"'value'",
")",
"if",
"name",
"==",
"'bool'",
":",
"if",
"length",
"is",
"not",
"None",
":",
"raise",
"ValueError",
"(",
"\"You can't specify a length with bool tokens - they are always one bit.\"",
")",
"length",
"=",
"1",
"if",
"length",
"is",
"None",
"and",
"name",
"not",
"in",
"(",
"'se'",
",",
"'ue'",
",",
"'sie'",
",",
"'uie'",
")",
":",
"stretchy_token",
"=",
"True",
"if",
"length",
"is",
"not",
"None",
":",
"# Try converting length to int, otherwise check it's a key.",
"try",
":",
"length",
"=",
"int",
"(",
"length",
")",
"if",
"length",
"<",
"0",
":",
"raise",
"Error",
"# For the 'bytes' token convert length to bits.",
"if",
"name",
"==",
"'bytes'",
":",
"length",
"*=",
"8",
"except",
"Error",
":",
"raise",
"ValueError",
"(",
"\"Can't read a token with a negative length.\"",
")",
"except",
"ValueError",
":",
"if",
"not",
"keys",
"or",
"length",
"not",
"in",
"keys",
":",
"raise",
"ValueError",
"(",
"\"Don't understand length '{0}' of token.\"",
".",
"format",
"(",
"length",
")",
")",
"ret_vals",
".",
"append",
"(",
"[",
"name",
",",
"length",
",",
"value",
"]",
")",
"# This multiplies by the multiplicative factor, but this means that",
"# we can't allow keyword values as multipliers (e.g. n*uint:8).",
"# The only way to do this would be to return the factor in some fashion",
"# (we can't use the key's value here as it would mean that we couldn't",
"# sensibly continue to cache the function's results. (TODO).",
"return_values",
".",
"extend",
"(",
"ret_vals",
"*",
"factor",
")",
"return_values",
"=",
"[",
"tuple",
"(",
"x",
")",
"for",
"x",
"in",
"return_values",
"]",
"if",
"len",
"(",
"token_cache",
")",
"<",
"CACHE_SIZE",
":",
"token_cache",
"[",
"token_key",
"]",
"=",
"stretchy_token",
",",
"return_values",
"return",
"stretchy_token",
",",
"return_values"
] | Divide the format string into tokens and parse them.
Return stretchy token and list of [initialiser, length, value]
initialiser is one of: hex, oct, bin, uint, int, se, ue, 0x, 0o, 0b etc.
length is None if not known, as is value.
If the token is in the keyword dictionary (keys) then it counts as a
special case and isn't messed with.
tokens must be of the form: [factor*][initialiser][:][length][=value] | [
"Divide",
"the",
"format",
"string",
"into",
"tokens",
"and",
"parse",
"them",
"."
] | ab40ae7f0b43fe223a39b63cbc0529b09f3ef653 | https://github.com/scott-griffiths/bitstring/blob/ab40ae7f0b43fe223a39b63cbc0529b09f3ef653/bitstring.py#L534-L633 | train |
scott-griffiths/bitstring | bitstring.py | expand_brackets | def expand_brackets(s):
"""Remove whitespace and expand all brackets."""
s = ''.join(s.split())
while True:
start = s.find('(')
if start == -1:
break
count = 1 # Number of hanging open brackets
p = start + 1
while p < len(s):
if s[p] == '(':
count += 1
if s[p] == ')':
count -= 1
if not count:
break
p += 1
if count:
raise ValueError("Unbalanced parenthesis in '{0}'.".format(s))
if start == 0 or s[start - 1] != '*':
s = s[0:start] + s[start + 1:p] + s[p + 1:]
else:
m = BRACKET_RE.search(s)
if m:
factor = int(m.group('factor'))
matchstart = m.start('factor')
s = s[0:matchstart] + (factor - 1) * (s[start + 1:p] + ',') + s[start + 1:p] + s[p + 1:]
else:
raise ValueError("Failed to parse '{0}'.".format(s))
return s | python | def expand_brackets(s):
"""Remove whitespace and expand all brackets."""
s = ''.join(s.split())
while True:
start = s.find('(')
if start == -1:
break
count = 1 # Number of hanging open brackets
p = start + 1
while p < len(s):
if s[p] == '(':
count += 1
if s[p] == ')':
count -= 1
if not count:
break
p += 1
if count:
raise ValueError("Unbalanced parenthesis in '{0}'.".format(s))
if start == 0 or s[start - 1] != '*':
s = s[0:start] + s[start + 1:p] + s[p + 1:]
else:
m = BRACKET_RE.search(s)
if m:
factor = int(m.group('factor'))
matchstart = m.start('factor')
s = s[0:matchstart] + (factor - 1) * (s[start + 1:p] + ',') + s[start + 1:p] + s[p + 1:]
else:
raise ValueError("Failed to parse '{0}'.".format(s))
return s | [
"def",
"expand_brackets",
"(",
"s",
")",
":",
"s",
"=",
"''",
".",
"join",
"(",
"s",
".",
"split",
"(",
")",
")",
"while",
"True",
":",
"start",
"=",
"s",
".",
"find",
"(",
"'('",
")",
"if",
"start",
"==",
"-",
"1",
":",
"break",
"count",
"=",
"1",
"# Number of hanging open brackets",
"p",
"=",
"start",
"+",
"1",
"while",
"p",
"<",
"len",
"(",
"s",
")",
":",
"if",
"s",
"[",
"p",
"]",
"==",
"'('",
":",
"count",
"+=",
"1",
"if",
"s",
"[",
"p",
"]",
"==",
"')'",
":",
"count",
"-=",
"1",
"if",
"not",
"count",
":",
"break",
"p",
"+=",
"1",
"if",
"count",
":",
"raise",
"ValueError",
"(",
"\"Unbalanced parenthesis in '{0}'.\"",
".",
"format",
"(",
"s",
")",
")",
"if",
"start",
"==",
"0",
"or",
"s",
"[",
"start",
"-",
"1",
"]",
"!=",
"'*'",
":",
"s",
"=",
"s",
"[",
"0",
":",
"start",
"]",
"+",
"s",
"[",
"start",
"+",
"1",
":",
"p",
"]",
"+",
"s",
"[",
"p",
"+",
"1",
":",
"]",
"else",
":",
"m",
"=",
"BRACKET_RE",
".",
"search",
"(",
"s",
")",
"if",
"m",
":",
"factor",
"=",
"int",
"(",
"m",
".",
"group",
"(",
"'factor'",
")",
")",
"matchstart",
"=",
"m",
".",
"start",
"(",
"'factor'",
")",
"s",
"=",
"s",
"[",
"0",
":",
"matchstart",
"]",
"+",
"(",
"factor",
"-",
"1",
")",
"*",
"(",
"s",
"[",
"start",
"+",
"1",
":",
"p",
"]",
"+",
"','",
")",
"+",
"s",
"[",
"start",
"+",
"1",
":",
"p",
"]",
"+",
"s",
"[",
"p",
"+",
"1",
":",
"]",
"else",
":",
"raise",
"ValueError",
"(",
"\"Failed to parse '{0}'.\"",
".",
"format",
"(",
"s",
")",
")",
"return",
"s"
] | Remove whitespace and expand all brackets. | [
"Remove",
"whitespace",
"and",
"expand",
"all",
"brackets",
"."
] | ab40ae7f0b43fe223a39b63cbc0529b09f3ef653 | https://github.com/scott-griffiths/bitstring/blob/ab40ae7f0b43fe223a39b63cbc0529b09f3ef653/bitstring.py#L638-L667 | train |
scott-griffiths/bitstring | bitstring.py | pack | def pack(fmt, *values, **kwargs):
"""Pack the values according to the format string and return a new BitStream.
fmt -- A single string or a list of strings with comma separated tokens
describing how to create the BitStream.
values -- Zero or more values to pack according to the format.
kwargs -- A dictionary or keyword-value pairs - the keywords used in the
format string will be replaced with their given value.
Token examples: 'int:12' : 12 bits as a signed integer
'uint:8' : 8 bits as an unsigned integer
'float:64' : 8 bytes as a big-endian float
'intbe:16' : 2 bytes as a big-endian signed integer
'uintbe:16' : 2 bytes as a big-endian unsigned integer
'intle:32' : 4 bytes as a little-endian signed integer
'uintle:32' : 4 bytes as a little-endian unsigned integer
'floatle:64': 8 bytes as a little-endian float
'intne:24' : 3 bytes as a native-endian signed integer
'uintne:24' : 3 bytes as a native-endian unsigned integer
'floatne:32': 4 bytes as a native-endian float
'hex:80' : 80 bits as a hex string
'oct:9' : 9 bits as an octal string
'bin:1' : single bit binary string
'ue' / 'uie': next bits as unsigned exp-Golomb code
'se' / 'sie': next bits as signed exp-Golomb code
'bits:5' : 5 bits as a bitstring object
'bytes:10' : 10 bytes as a bytes object
'bool' : 1 bit as a bool
'pad:3' : 3 zero bits as padding
>>> s = pack('uint:12, bits', 100, '0xffe')
>>> t = pack(['bits', 'bin:3'], s, '111')
>>> u = pack('uint:8=a, uint:8=b, uint:55=a', a=6, b=44)
"""
tokens = []
if isinstance(fmt, basestring):
fmt = [fmt]
try:
for f_item in fmt:
_, tkns = tokenparser(f_item, tuple(sorted(kwargs.keys())))
tokens.extend(tkns)
except ValueError as e:
raise CreationError(*e.args)
value_iter = iter(values)
s = BitStream()
try:
for name, length, value in tokens:
# If the value is in the kwd dictionary then it takes precedence.
if value in kwargs:
value = kwargs[value]
# If the length is in the kwd dictionary then use that too.
if length in kwargs:
length = kwargs[length]
# Also if we just have a dictionary name then we want to use it
if name in kwargs and length is None and value is None:
s.append(kwargs[name])
continue
if length is not None:
length = int(length)
if value is None and name != 'pad':
# Take the next value from the ones provided
value = next(value_iter)
s._append(BitStream._init_with_token(name, length, value))
except StopIteration:
raise CreationError("Not enough parameters present to pack according to the "
"format. {0} values are needed.", len(tokens))
try:
next(value_iter)
except StopIteration:
# Good, we've used up all the *values.
return s
raise CreationError("Too many parameters present to pack according to the format.") | python | def pack(fmt, *values, **kwargs):
"""Pack the values according to the format string and return a new BitStream.
fmt -- A single string or a list of strings with comma separated tokens
describing how to create the BitStream.
values -- Zero or more values to pack according to the format.
kwargs -- A dictionary or keyword-value pairs - the keywords used in the
format string will be replaced with their given value.
Token examples: 'int:12' : 12 bits as a signed integer
'uint:8' : 8 bits as an unsigned integer
'float:64' : 8 bytes as a big-endian float
'intbe:16' : 2 bytes as a big-endian signed integer
'uintbe:16' : 2 bytes as a big-endian unsigned integer
'intle:32' : 4 bytes as a little-endian signed integer
'uintle:32' : 4 bytes as a little-endian unsigned integer
'floatle:64': 8 bytes as a little-endian float
'intne:24' : 3 bytes as a native-endian signed integer
'uintne:24' : 3 bytes as a native-endian unsigned integer
'floatne:32': 4 bytes as a native-endian float
'hex:80' : 80 bits as a hex string
'oct:9' : 9 bits as an octal string
'bin:1' : single bit binary string
'ue' / 'uie': next bits as unsigned exp-Golomb code
'se' / 'sie': next bits as signed exp-Golomb code
'bits:5' : 5 bits as a bitstring object
'bytes:10' : 10 bytes as a bytes object
'bool' : 1 bit as a bool
'pad:3' : 3 zero bits as padding
>>> s = pack('uint:12, bits', 100, '0xffe')
>>> t = pack(['bits', 'bin:3'], s, '111')
>>> u = pack('uint:8=a, uint:8=b, uint:55=a', a=6, b=44)
"""
tokens = []
if isinstance(fmt, basestring):
fmt = [fmt]
try:
for f_item in fmt:
_, tkns = tokenparser(f_item, tuple(sorted(kwargs.keys())))
tokens.extend(tkns)
except ValueError as e:
raise CreationError(*e.args)
value_iter = iter(values)
s = BitStream()
try:
for name, length, value in tokens:
# If the value is in the kwd dictionary then it takes precedence.
if value in kwargs:
value = kwargs[value]
# If the length is in the kwd dictionary then use that too.
if length in kwargs:
length = kwargs[length]
# Also if we just have a dictionary name then we want to use it
if name in kwargs and length is None and value is None:
s.append(kwargs[name])
continue
if length is not None:
length = int(length)
if value is None and name != 'pad':
# Take the next value from the ones provided
value = next(value_iter)
s._append(BitStream._init_with_token(name, length, value))
except StopIteration:
raise CreationError("Not enough parameters present to pack according to the "
"format. {0} values are needed.", len(tokens))
try:
next(value_iter)
except StopIteration:
# Good, we've used up all the *values.
return s
raise CreationError("Too many parameters present to pack according to the format.") | [
"def",
"pack",
"(",
"fmt",
",",
"*",
"values",
",",
"*",
"*",
"kwargs",
")",
":",
"tokens",
"=",
"[",
"]",
"if",
"isinstance",
"(",
"fmt",
",",
"basestring",
")",
":",
"fmt",
"=",
"[",
"fmt",
"]",
"try",
":",
"for",
"f_item",
"in",
"fmt",
":",
"_",
",",
"tkns",
"=",
"tokenparser",
"(",
"f_item",
",",
"tuple",
"(",
"sorted",
"(",
"kwargs",
".",
"keys",
"(",
")",
")",
")",
")",
"tokens",
".",
"extend",
"(",
"tkns",
")",
"except",
"ValueError",
"as",
"e",
":",
"raise",
"CreationError",
"(",
"*",
"e",
".",
"args",
")",
"value_iter",
"=",
"iter",
"(",
"values",
")",
"s",
"=",
"BitStream",
"(",
")",
"try",
":",
"for",
"name",
",",
"length",
",",
"value",
"in",
"tokens",
":",
"# If the value is in the kwd dictionary then it takes precedence.",
"if",
"value",
"in",
"kwargs",
":",
"value",
"=",
"kwargs",
"[",
"value",
"]",
"# If the length is in the kwd dictionary then use that too.",
"if",
"length",
"in",
"kwargs",
":",
"length",
"=",
"kwargs",
"[",
"length",
"]",
"# Also if we just have a dictionary name then we want to use it",
"if",
"name",
"in",
"kwargs",
"and",
"length",
"is",
"None",
"and",
"value",
"is",
"None",
":",
"s",
".",
"append",
"(",
"kwargs",
"[",
"name",
"]",
")",
"continue",
"if",
"length",
"is",
"not",
"None",
":",
"length",
"=",
"int",
"(",
"length",
")",
"if",
"value",
"is",
"None",
"and",
"name",
"!=",
"'pad'",
":",
"# Take the next value from the ones provided",
"value",
"=",
"next",
"(",
"value_iter",
")",
"s",
".",
"_append",
"(",
"BitStream",
".",
"_init_with_token",
"(",
"name",
",",
"length",
",",
"value",
")",
")",
"except",
"StopIteration",
":",
"raise",
"CreationError",
"(",
"\"Not enough parameters present to pack according to the \"",
"\"format. {0} values are needed.\"",
",",
"len",
"(",
"tokens",
")",
")",
"try",
":",
"next",
"(",
"value_iter",
")",
"except",
"StopIteration",
":",
"# Good, we've used up all the *values.",
"return",
"s",
"raise",
"CreationError",
"(",
"\"Too many parameters present to pack according to the format.\"",
")"
] | Pack the values according to the format string and return a new BitStream.
fmt -- A single string or a list of strings with comma separated tokens
describing how to create the BitStream.
values -- Zero or more values to pack according to the format.
kwargs -- A dictionary or keyword-value pairs - the keywords used in the
format string will be replaced with their given value.
Token examples: 'int:12' : 12 bits as a signed integer
'uint:8' : 8 bits as an unsigned integer
'float:64' : 8 bytes as a big-endian float
'intbe:16' : 2 bytes as a big-endian signed integer
'uintbe:16' : 2 bytes as a big-endian unsigned integer
'intle:32' : 4 bytes as a little-endian signed integer
'uintle:32' : 4 bytes as a little-endian unsigned integer
'floatle:64': 8 bytes as a little-endian float
'intne:24' : 3 bytes as a native-endian signed integer
'uintne:24' : 3 bytes as a native-endian unsigned integer
'floatne:32': 4 bytes as a native-endian float
'hex:80' : 80 bits as a hex string
'oct:9' : 9 bits as an octal string
'bin:1' : single bit binary string
'ue' / 'uie': next bits as unsigned exp-Golomb code
'se' / 'sie': next bits as signed exp-Golomb code
'bits:5' : 5 bits as a bitstring object
'bytes:10' : 10 bytes as a bytes object
'bool' : 1 bit as a bool
'pad:3' : 3 zero bits as padding
>>> s = pack('uint:12, bits', 100, '0xffe')
>>> t = pack(['bits', 'bin:3'], s, '111')
>>> u = pack('uint:8=a, uint:8=b, uint:55=a', a=6, b=44) | [
"Pack",
"the",
"values",
"according",
"to",
"the",
"format",
"string",
"and",
"return",
"a",
"new",
"BitStream",
"."
] | ab40ae7f0b43fe223a39b63cbc0529b09f3ef653 | https://github.com/scott-griffiths/bitstring/blob/ab40ae7f0b43fe223a39b63cbc0529b09f3ef653/bitstring.py#L4161-L4233 | train |
scott-griffiths/bitstring | bitstring.py | ConstByteStore.getbyteslice | def getbyteslice(self, start, end):
"""Direct access to byte data."""
c = self._rawarray[start:end]
return c | python | def getbyteslice(self, start, end):
"""Direct access to byte data."""
c = self._rawarray[start:end]
return c | [
"def",
"getbyteslice",
"(",
"self",
",",
"start",
",",
"end",
")",
":",
"c",
"=",
"self",
".",
"_rawarray",
"[",
"start",
":",
"end",
"]",
"return",
"c"
] | Direct access to byte data. | [
"Direct",
"access",
"to",
"byte",
"data",
"."
] | ab40ae7f0b43fe223a39b63cbc0529b09f3ef653 | https://github.com/scott-griffiths/bitstring/blob/ab40ae7f0b43fe223a39b63cbc0529b09f3ef653/bitstring.py#L157-L160 | train |