Reformat code using PyCharm
This commit is contained in:
parent
86e0fbe556
commit
f1594cfdde
33 changed files with 675 additions and 666 deletions
|
@ -1,20 +1,22 @@
|
||||||
from tls_client import Session
|
|
||||||
from forefront.mail import Mail
|
|
||||||
from time import time, sleep
|
|
||||||
from re import match
|
|
||||||
from forefront.typing import ForeFrontResponse
|
|
||||||
from uuid import uuid4
|
|
||||||
from requests import post
|
|
||||||
from json import loads
|
from json import loads
|
||||||
|
from re import match
|
||||||
|
from time import time, sleep
|
||||||
|
from uuid import uuid4
|
||||||
|
|
||||||
|
from requests import post
|
||||||
|
from tls_client import Session
|
||||||
|
|
||||||
|
from forefront.mail import Mail
|
||||||
|
from forefront.typing import ForeFrontResponse
|
||||||
|
|
||||||
|
|
||||||
class Account:
|
class Account:
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def create(proxy = None, logging = False):
|
def create(proxy=None, logging=False):
|
||||||
|
|
||||||
proxies = {
|
proxies = {
|
||||||
'http': 'http://' + proxy,
|
'http': 'http://' + proxy,
|
||||||
'https': 'http://' + proxy } if proxy else False
|
'https': 'http://' + proxy} if proxy else False
|
||||||
|
|
||||||
start = time()
|
start = time()
|
||||||
|
|
||||||
|
@ -22,17 +24,17 @@ class Account:
|
||||||
mail_token = None
|
mail_token = None
|
||||||
mail_adress = mail.get_mail()
|
mail_adress = mail.get_mail()
|
||||||
|
|
||||||
#print(mail_adress)
|
# print(mail_adress)
|
||||||
|
|
||||||
client = Session(client_identifier='chrome110')
|
client = Session(client_identifier='chrome110')
|
||||||
client.proxies = proxies
|
client.proxies = proxies
|
||||||
client.headers = {
|
client.headers = {
|
||||||
"origin": "https://accounts.forefront.ai",
|
"origin": "https://accounts.forefront.ai",
|
||||||
"user-agent" : "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36",
|
"user-agent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36",
|
||||||
}
|
}
|
||||||
|
|
||||||
response = client.post('https://clerk.forefront.ai/v1/client/sign_ups?_clerk_js_version=4.32.6',
|
response = client.post('https://clerk.forefront.ai/v1/client/sign_ups?_clerk_js_version=4.32.6',
|
||||||
data = {
|
data={
|
||||||
"email_address": mail_adress
|
"email_address": mail_adress
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
@ -40,9 +42,10 @@ class Account:
|
||||||
trace_token = response.json()['response']['id']
|
trace_token = response.json()['response']['id']
|
||||||
if logging: print(trace_token)
|
if logging: print(trace_token)
|
||||||
|
|
||||||
response = client.post(f"https://clerk.forefront.ai/v1/client/sign_ups/{trace_token}/prepare_verification?_clerk_js_version=4.32.6",
|
response = client.post(
|
||||||
data = {
|
f"https://clerk.forefront.ai/v1/client/sign_ups/{trace_token}/prepare_verification?_clerk_js_version=4.32.6",
|
||||||
"strategy" : "email_code",
|
data={
|
||||||
|
"strategy": "email_code",
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -62,7 +65,9 @@ class Account:
|
||||||
|
|
||||||
if logging: print(mail_token)
|
if logging: print(mail_token)
|
||||||
|
|
||||||
response = client.post(f'https://clerk.forefront.ai/v1/client/sign_ups/{trace_token}/attempt_verification?_clerk_js_version=4.38.4', data = {
|
response = client.post(
|
||||||
|
f'https://clerk.forefront.ai/v1/client/sign_ups/{trace_token}/attempt_verification?_clerk_js_version=4.38.4',
|
||||||
|
data={
|
||||||
'code': mail_token,
|
'code': mail_token,
|
||||||
'strategy': 'email_code'
|
'strategy': 'email_code'
|
||||||
})
|
})
|
||||||
|
@ -82,42 +87,42 @@ class Account:
|
||||||
class StreamingCompletion:
|
class StreamingCompletion:
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def create(
|
def create(
|
||||||
token = None,
|
token=None,
|
||||||
chatId = None,
|
chatId=None,
|
||||||
prompt = '',
|
prompt='',
|
||||||
actionType = 'new',
|
actionType='new',
|
||||||
defaultPersona = '607e41fe-95be-497e-8e97-010a59b2e2c0', # default
|
defaultPersona='607e41fe-95be-497e-8e97-010a59b2e2c0', # default
|
||||||
model = 'gpt-4') -> ForeFrontResponse:
|
model='gpt-4') -> ForeFrontResponse:
|
||||||
|
|
||||||
if not token: raise Exception('Token is required!')
|
if not token: raise Exception('Token is required!')
|
||||||
if not chatId: chatId = str(uuid4())
|
if not chatId: chatId = str(uuid4())
|
||||||
|
|
||||||
headers = {
|
headers = {
|
||||||
'authority' : 'chat-server.tenant-forefront-default.knative.chi.coreweave.com',
|
'authority': 'chat-server.tenant-forefront-default.knative.chi.coreweave.com',
|
||||||
'accept' : '*/*',
|
'accept': '*/*',
|
||||||
'accept-language' : 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3',
|
'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3',
|
||||||
'authorization' : 'Bearer ' + token,
|
'authorization': 'Bearer ' + token,
|
||||||
'cache-control' : 'no-cache',
|
'cache-control': 'no-cache',
|
||||||
'content-type' : 'application/json',
|
'content-type': 'application/json',
|
||||||
'origin' : 'https://chat.forefront.ai',
|
'origin': 'https://chat.forefront.ai',
|
||||||
'pragma' : 'no-cache',
|
'pragma': 'no-cache',
|
||||||
'referer' : 'https://chat.forefront.ai/',
|
'referer': 'https://chat.forefront.ai/',
|
||||||
'sec-ch-ua' : '"Chromium";v="112", "Google Chrome";v="112", "Not:A-Brand";v="99"',
|
'sec-ch-ua': '"Chromium";v="112", "Google Chrome";v="112", "Not:A-Brand";v="99"',
|
||||||
'sec-ch-ua-mobile' : '?0',
|
'sec-ch-ua-mobile': '?0',
|
||||||
'sec-ch-ua-platform': '"macOS"',
|
'sec-ch-ua-platform': '"macOS"',
|
||||||
'sec-fetch-dest' : 'empty',
|
'sec-fetch-dest': 'empty',
|
||||||
'sec-fetch-mode' : 'cors',
|
'sec-fetch-mode': 'cors',
|
||||||
'sec-fetch-site' : 'cross-site',
|
'sec-fetch-site': 'cross-site',
|
||||||
'user-agent' : 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36',
|
'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36',
|
||||||
}
|
}
|
||||||
|
|
||||||
json_data = {
|
json_data = {
|
||||||
'text' : prompt,
|
'text': prompt,
|
||||||
'action' : actionType,
|
'action': actionType,
|
||||||
'parentId' : chatId,
|
'parentId': chatId,
|
||||||
'workspaceId' : chatId,
|
'workspaceId': chatId,
|
||||||
'messagePersona' : defaultPersona,
|
'messagePersona': defaultPersona,
|
||||||
'model' : model
|
'model': model
|
||||||
}
|
}
|
||||||
|
|
||||||
for chunk in post('https://chat-server.tenant-forefront-default.knative.chi.coreweave.com/chat',
|
for chunk in post('https://chat-server.tenant-forefront-default.knative.chi.coreweave.com/chat',
|
||||||
|
@ -129,19 +134,19 @@ class StreamingCompletion:
|
||||||
|
|
||||||
if token != None:
|
if token != None:
|
||||||
yield ForeFrontResponse({
|
yield ForeFrontResponse({
|
||||||
'id' : chatId,
|
'id': chatId,
|
||||||
'object' : 'text_completion',
|
'object': 'text_completion',
|
||||||
'created': int(time()),
|
'created': int(time()),
|
||||||
'model' : model,
|
'model': model,
|
||||||
'choices': [{
|
'choices': [{
|
||||||
'text' : token,
|
'text': token,
|
||||||
'index' : 0,
|
'index': 0,
|
||||||
'logprobs' : None,
|
'logprobs': None,
|
||||||
'finish_reason' : 'stop'
|
'finish_reason': 'stop'
|
||||||
}],
|
}],
|
||||||
'usage': {
|
'usage': {
|
||||||
'prompt_tokens' : len(prompt),
|
'prompt_tokens': len(prompt),
|
||||||
'completion_tokens' : len(token),
|
'completion_tokens': len(token),
|
||||||
'total_tokens' : len(prompt) + len(token)
|
'total_tokens': len(prompt) + len(token)
|
||||||
}
|
}
|
||||||
})
|
})
|
|
@ -1,6 +1,8 @@
|
||||||
from requests import Session
|
|
||||||
from string import ascii_letters
|
|
||||||
from random import choices
|
from random import choices
|
||||||
|
from string import ascii_letters
|
||||||
|
|
||||||
|
from requests import Session
|
||||||
|
|
||||||
|
|
||||||
class Mail:
|
class Mail:
|
||||||
def __init__(self, proxies: dict = None) -> None:
|
def __init__(self, proxies: dict = None) -> None:
|
||||||
|
@ -27,12 +29,12 @@ class Mail:
|
||||||
def get_mail(self) -> str:
|
def get_mail(self) -> str:
|
||||||
token = ''.join(choices(ascii_letters, k=14)).lower()
|
token = ''.join(choices(ascii_letters, k=14)).lower()
|
||||||
init = self.client.post("https://api.mail.tm/accounts", json={
|
init = self.client.post("https://api.mail.tm/accounts", json={
|
||||||
"address" : f"{token}@bugfoo.com",
|
"address": f"{token}@bugfoo.com",
|
||||||
"password": token
|
"password": token
|
||||||
})
|
})
|
||||||
|
|
||||||
if init.status_code == 201:
|
if init.status_code == 201:
|
||||||
resp = self.client.post("https://api.mail.tm/token", json = {
|
resp = self.client.post("https://api.mail.tm/token", json={
|
||||||
**init.json(),
|
**init.json(),
|
||||||
"password": token
|
"password": token
|
||||||
})
|
})
|
||||||
|
@ -52,4 +54,3 @@ class Mail:
|
||||||
|
|
||||||
def get_message_content(self, message_id: str):
|
def get_message_content(self, message_id: str):
|
||||||
return self.get_message(message_id)["text"]
|
return self.get_message(message_id)["text"]
|
||||||
|
|
||||||
|
|
|
@ -24,7 +24,6 @@ class ForeFrontResponse:
|
||||||
return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>'''
|
return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>'''
|
||||||
|
|
||||||
def __init__(self, response_dict: dict) -> None:
|
def __init__(self, response_dict: dict) -> None:
|
||||||
|
|
||||||
self.response_dict = response_dict
|
self.response_dict = response_dict
|
||||||
self.id = response_dict['id']
|
self.id = response_dict['id']
|
||||||
self.object = response_dict['object']
|
self.object = response_dict['object']
|
||||||
|
|
|
@ -1,19 +1,17 @@
|
||||||
from urllib.parse import quote
|
|
||||||
from time import time
|
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from queue import Queue, Empty
|
from queue import Queue, Empty
|
||||||
from threading import Thread
|
from threading import Thread
|
||||||
from re import findall
|
from time import time
|
||||||
|
from urllib.parse import quote
|
||||||
|
|
||||||
from curl_cffi.requests import post
|
from curl_cffi.requests import post
|
||||||
|
|
||||||
cf_clearance = ''
|
cf_clearance = ''
|
||||||
user_agent = 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36'
|
user_agent = 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36'
|
||||||
|
|
||||||
|
|
||||||
class PhindResponse:
|
class PhindResponse:
|
||||||
|
|
||||||
class Completion:
|
class Completion:
|
||||||
|
|
||||||
class Choices:
|
class Choices:
|
||||||
def __init__(self, choice: dict) -> None:
|
def __init__(self, choice: dict) -> None:
|
||||||
self.text = choice['text']
|
self.text = choice['text']
|
||||||
|
@ -38,7 +36,6 @@ class PhindResponse:
|
||||||
return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>'''
|
return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>'''
|
||||||
|
|
||||||
def __init__(self, response_dict: dict) -> None:
|
def __init__(self, response_dict: dict) -> None:
|
||||||
|
|
||||||
self.response_dict = response_dict
|
self.response_dict = response_dict
|
||||||
self.id = response_dict['id']
|
self.id = response_dict['id']
|
||||||
self.object = response_dict['object']
|
self.object = response_dict['object']
|
||||||
|
@ -55,7 +52,7 @@ class Search:
|
||||||
def create(prompt: str, actualSearch: bool = True, language: str = 'en') -> dict: # None = no search
|
def create(prompt: str, actualSearch: bool = True, language: str = 'en') -> dict: # None = no search
|
||||||
if user_agent == '':
|
if user_agent == '':
|
||||||
raise ValueError('user_agent must be set, refer to documentation')
|
raise ValueError('user_agent must be set, refer to documentation')
|
||||||
if cf_clearance == '' :
|
if cf_clearance == '':
|
||||||
raise ValueError('cf_clearance must be set, refer to documentation')
|
raise ValueError('cf_clearance must be set, refer to documentation')
|
||||||
|
|
||||||
if not actualSearch:
|
if not actualSearch:
|
||||||
|
@ -92,7 +89,7 @@ class Search:
|
||||||
'user-agent': user_agent
|
'user-agent': user_agent
|
||||||
}
|
}
|
||||||
|
|
||||||
return post('https://www.phind.com/api/bing/search', headers = headers, json = {
|
return post('https://www.phind.com/api/bing/search', headers=headers, json={
|
||||||
'q': prompt,
|
'q': prompt,
|
||||||
'userRankList': {},
|
'userRankList': {},
|
||||||
'browserLanguage': language}).json()['rawBingResults']
|
'browserLanguage': language}).json()['rawBingResults']
|
||||||
|
@ -100,7 +97,7 @@ class Search:
|
||||||
|
|
||||||
class Completion:
|
class Completion:
|
||||||
def create(
|
def create(
|
||||||
model = 'gpt-4',
|
model='gpt-4',
|
||||||
prompt: str = '',
|
prompt: str = '',
|
||||||
results: dict = None,
|
results: dict = None,
|
||||||
creative: bool = False,
|
creative: bool = False,
|
||||||
|
@ -108,31 +105,31 @@ class Completion:
|
||||||
codeContext: str = '',
|
codeContext: str = '',
|
||||||
language: str = 'en') -> PhindResponse:
|
language: str = 'en') -> PhindResponse:
|
||||||
|
|
||||||
if user_agent == '' :
|
if user_agent == '':
|
||||||
raise ValueError('user_agent must be set, refer to documentation')
|
raise ValueError('user_agent must be set, refer to documentation')
|
||||||
|
|
||||||
if cf_clearance == '' :
|
if cf_clearance == '':
|
||||||
raise ValueError('cf_clearance must be set, refer to documentation')
|
raise ValueError('cf_clearance must be set, refer to documentation')
|
||||||
|
|
||||||
if results is None:
|
if results is None:
|
||||||
results = Search.create(prompt, actualSearch = True)
|
results = Search.create(prompt, actualSearch=True)
|
||||||
|
|
||||||
if len(codeContext) > 2999:
|
if len(codeContext) > 2999:
|
||||||
raise ValueError('codeContext must be less than 3000 characters')
|
raise ValueError('codeContext must be less than 3000 characters')
|
||||||
|
|
||||||
models = {
|
models = {
|
||||||
'gpt-4' : 'expert',
|
'gpt-4': 'expert',
|
||||||
'gpt-3.5-turbo' : 'intermediate',
|
'gpt-3.5-turbo': 'intermediate',
|
||||||
'gpt-3.5': 'intermediate',
|
'gpt-3.5': 'intermediate',
|
||||||
}
|
}
|
||||||
|
|
||||||
json_data = {
|
json_data = {
|
||||||
'question' : prompt,
|
'question': prompt,
|
||||||
'bingResults' : results, #response.json()['rawBingResults'],
|
'bingResults': results, # response.json()['rawBingResults'],
|
||||||
'codeContext' : codeContext,
|
'codeContext': codeContext,
|
||||||
'options': {
|
'options': {
|
||||||
'skill' : models[model],
|
'skill': models[model],
|
||||||
'date' : datetime.now().strftime("%d/%m/%Y"),
|
'date': datetime.now().strftime("%d/%m/%Y"),
|
||||||
'language': language,
|
'language': language,
|
||||||
'detailed': detailed,
|
'detailed': detailed,
|
||||||
'creative': creative
|
'creative': creative
|
||||||
|
@ -157,25 +154,26 @@ class Completion:
|
||||||
}
|
}
|
||||||
|
|
||||||
completion = ''
|
completion = ''
|
||||||
response = post('https://www.phind.com/api/infer/answer', headers = headers, json = json_data, timeout=99999, impersonate='chrome110')
|
response = post('https://www.phind.com/api/infer/answer', headers=headers, json=json_data, timeout=99999,
|
||||||
|
impersonate='chrome110')
|
||||||
for line in response.text.split('\r\n\r\n'):
|
for line in response.text.split('\r\n\r\n'):
|
||||||
completion += (line.replace('data: ', ''))
|
completion += (line.replace('data: ', ''))
|
||||||
|
|
||||||
return PhindResponse({
|
return PhindResponse({
|
||||||
'id' : f'cmpl-1337-{int(time())}',
|
'id': f'cmpl-1337-{int(time())}',
|
||||||
'object' : 'text_completion',
|
'object': 'text_completion',
|
||||||
'created': int(time()),
|
'created': int(time()),
|
||||||
'model' : models[model],
|
'model': models[model],
|
||||||
'choices': [{
|
'choices': [{
|
||||||
'text' : completion,
|
'text': completion,
|
||||||
'index' : 0,
|
'index': 0,
|
||||||
'logprobs' : None,
|
'logprobs': None,
|
||||||
'finish_reason' : 'stop'
|
'finish_reason': 'stop'
|
||||||
}],
|
}],
|
||||||
'usage': {
|
'usage': {
|
||||||
'prompt_tokens' : len(prompt),
|
'prompt_tokens': len(prompt),
|
||||||
'completion_tokens' : len(completion),
|
'completion_tokens': len(completion),
|
||||||
'total_tokens' : len(prompt) + len(completion)
|
'total_tokens': len(prompt) + len(completion)
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
|
@ -187,18 +185,18 @@ class StreamingCompletion:
|
||||||
def request(model, prompt, results, creative, detailed, codeContext, language) -> None:
|
def request(model, prompt, results, creative, detailed, codeContext, language) -> None:
|
||||||
|
|
||||||
models = {
|
models = {
|
||||||
'gpt-4' : 'expert',
|
'gpt-4': 'expert',
|
||||||
'gpt-3.5-turbo' : 'intermediate',
|
'gpt-3.5-turbo': 'intermediate',
|
||||||
'gpt-3.5': 'intermediate',
|
'gpt-3.5': 'intermediate',
|
||||||
}
|
}
|
||||||
|
|
||||||
json_data = {
|
json_data = {
|
||||||
'question' : prompt,
|
'question': prompt,
|
||||||
'bingResults' : results,
|
'bingResults': results,
|
||||||
'codeContext' : codeContext,
|
'codeContext': codeContext,
|
||||||
'options': {
|
'options': {
|
||||||
'skill' : models[model],
|
'skill': models[model],
|
||||||
'date' : datetime.now().strftime("%d/%m/%Y"),
|
'date': datetime.now().strftime("%d/%m/%Y"),
|
||||||
'language': language,
|
'language': language,
|
||||||
'detailed': detailed,
|
'detailed': detailed,
|
||||||
'creative': creative
|
'creative': creative
|
||||||
|
@ -223,33 +221,33 @@ class StreamingCompletion:
|
||||||
}
|
}
|
||||||
|
|
||||||
response = post('https://www.phind.com/api/infer/answer',
|
response = post('https://www.phind.com/api/infer/answer',
|
||||||
headers = headers, json = json_data, timeout=99999, impersonate='chrome110', content_callback=StreamingCompletion.handle_stream_response)
|
headers=headers, json=json_data, timeout=99999, impersonate='chrome110',
|
||||||
|
content_callback=StreamingCompletion.handle_stream_response)
|
||||||
|
|
||||||
StreamingCompletion.stream_completed = True
|
StreamingCompletion.stream_completed = True
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def create(
|
def create(
|
||||||
model : str = 'gpt-4',
|
model: str = 'gpt-4',
|
||||||
prompt : str = '',
|
prompt: str = '',
|
||||||
results : dict = None,
|
results: dict = None,
|
||||||
creative : bool = False,
|
creative: bool = False,
|
||||||
detailed : bool = False,
|
detailed: bool = False,
|
||||||
codeContext : str = '',
|
codeContext: str = '',
|
||||||
language : str = 'en'):
|
language: str = 'en'):
|
||||||
|
|
||||||
if user_agent == '':
|
if user_agent == '':
|
||||||
raise ValueError('user_agent must be set, refer to documentation')
|
raise ValueError('user_agent must be set, refer to documentation')
|
||||||
if cf_clearance == '' :
|
if cf_clearance == '':
|
||||||
raise ValueError('cf_clearance must be set, refer to documentation')
|
raise ValueError('cf_clearance must be set, refer to documentation')
|
||||||
|
|
||||||
if results is None:
|
if results is None:
|
||||||
results = Search.create(prompt, actualSearch = True)
|
results = Search.create(prompt, actualSearch=True)
|
||||||
|
|
||||||
if len(codeContext) > 2999:
|
if len(codeContext) > 2999:
|
||||||
raise ValueError('codeContext must be less than 3000 characters')
|
raise ValueError('codeContext must be less than 3000 characters')
|
||||||
|
|
||||||
Thread(target = StreamingCompletion.request, args = [
|
Thread(target=StreamingCompletion.request, args=[
|
||||||
model, prompt, results, creative, detailed, codeContext, language]).start()
|
model, prompt, results, creative, detailed, codeContext, language]).start()
|
||||||
|
|
||||||
while StreamingCompletion.stream_completed != True or not StreamingCompletion.message_queue.empty():
|
while StreamingCompletion.stream_completed != True or not StreamingCompletion.message_queue.empty():
|
||||||
|
@ -266,20 +264,20 @@ class StreamingCompletion:
|
||||||
chunk = chunk.replace('data: ', '').replace('\r\n\r\n', '')
|
chunk = chunk.replace('data: ', '').replace('\r\n\r\n', '')
|
||||||
|
|
||||||
yield PhindResponse({
|
yield PhindResponse({
|
||||||
'id' : f'cmpl-1337-{int(time())}',
|
'id': f'cmpl-1337-{int(time())}',
|
||||||
'object' : 'text_completion',
|
'object': 'text_completion',
|
||||||
'created': int(time()),
|
'created': int(time()),
|
||||||
'model' : model,
|
'model': model,
|
||||||
'choices': [{
|
'choices': [{
|
||||||
'text' : chunk,
|
'text': chunk,
|
||||||
'index' : 0,
|
'index': 0,
|
||||||
'logprobs' : None,
|
'logprobs': None,
|
||||||
'finish_reason' : 'stop'
|
'finish_reason': 'stop'
|
||||||
}],
|
}],
|
||||||
'usage': {
|
'usage': {
|
||||||
'prompt_tokens' : len(prompt),
|
'prompt_tokens': len(prompt),
|
||||||
'completion_tokens' : len(chunk),
|
'completion_tokens': len(chunk),
|
||||||
'total_tokens' : len(prompt) + len(chunk)
|
'total_tokens': len(prompt) + len(chunk)
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
|
|
|
@ -384,7 +384,7 @@ class Client:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
# update info about response
|
# update info about response
|
||||||
message["text_new"] = message["text"][len(last_text) :]
|
message["text_new"] = message["text"][len(last_text):]
|
||||||
last_text = message["text"]
|
last_text = message["text"]
|
||||||
message_id = message["messageId"]
|
message_id = message["messageId"]
|
||||||
|
|
||||||
|
|
|
@ -5,7 +5,6 @@ token = forefront.Account.create(logging=True)
|
||||||
print(token)
|
print(token)
|
||||||
|
|
||||||
# get a response
|
# get a response
|
||||||
for response in forefront.StreamingCompletion.create(token = token,
|
for response in forefront.StreamingCompletion.create(token=token,
|
||||||
prompt = 'hello world', model='gpt-4'):
|
prompt='hello world', model='gpt-4'):
|
||||||
|
print(response.completion.choices[0].text, end='')
|
||||||
print(response.completion.choices[0].text, end = '')
|
|
||||||
|
|
|
@ -8,12 +8,13 @@ prompt = 'hello world'
|
||||||
|
|
||||||
# normal completion
|
# normal completion
|
||||||
result = phind.Completion.create(
|
result = phind.Completion.create(
|
||||||
model = 'gpt-4',
|
model='gpt-4',
|
||||||
prompt = prompt,
|
prompt=prompt,
|
||||||
results = phind.Search.create(prompt, actualSearch = False), # create search (set actualSearch to False to disable internet)
|
results=phind.Search.create(prompt, actualSearch=False),
|
||||||
creative = False,
|
# create search (set actualSearch to False to disable internet)
|
||||||
detailed = False,
|
creative=False,
|
||||||
codeContext = '') # up to 3000 chars of code
|
detailed=False,
|
||||||
|
codeContext='') # up to 3000 chars of code
|
||||||
|
|
||||||
print(result.completion.choices[0].text)
|
print(result.completion.choices[0].text)
|
||||||
|
|
||||||
|
@ -22,11 +23,12 @@ prompt = 'who won the quatar world cup'
|
||||||
# help needed: not getting newlines from the stream, please submit a PR if you know how to fix this
|
# help needed: not getting newlines from the stream, please submit a PR if you know how to fix this
|
||||||
# stream completion
|
# stream completion
|
||||||
for result in phind.StreamingCompletion.create(
|
for result in phind.StreamingCompletion.create(
|
||||||
model = 'gpt-4',
|
model='gpt-4',
|
||||||
prompt = prompt,
|
prompt=prompt,
|
||||||
results = phind.Search.create(prompt, actualSearch = True), # create search (set actualSearch to False to disable internet)
|
results=phind.Search.create(prompt, actualSearch=True),
|
||||||
creative = False,
|
# create search (set actualSearch to False to disable internet)
|
||||||
detailed = False,
|
creative=False,
|
||||||
codeContext = ''): # up to 3000 chars of code
|
detailed=False,
|
||||||
|
codeContext=''): # up to 3000 chars of code
|
||||||
|
|
||||||
print(result.completion.choices[0].text, end='', flush=True)
|
print(result.completion.choices[0].text, end='', flush=True)
|
|
@ -1,16 +1,16 @@
|
||||||
from requests import Session
|
|
||||||
from tls_client import Session as TLS
|
|
||||||
from json import dumps
|
|
||||||
from hashlib import md5
|
from hashlib import md5
|
||||||
from time import sleep
|
from json import dumps
|
||||||
from re import findall
|
from re import findall
|
||||||
from pypasser import reCaptchaV3
|
|
||||||
from quora import extract_formkey
|
from tls_client import Session as TLS
|
||||||
from quora.mail import Emailnator
|
|
||||||
from twocaptcha import TwoCaptcha
|
from twocaptcha import TwoCaptcha
|
||||||
|
|
||||||
|
from quora import extract_formkey
|
||||||
|
from quora.mail import Emailnator
|
||||||
|
|
||||||
solver = TwoCaptcha('72747bf24a9d89b4dcc1b24875efd358')
|
solver = TwoCaptcha('72747bf24a9d89b4dcc1b24875efd358')
|
||||||
|
|
||||||
|
|
||||||
class Account:
|
class Account:
|
||||||
def create(proxy: None or str = None, logging: bool = False, enable_bot_creation: bool = False):
|
def create(proxy: None or str = None, logging: bool = False, enable_bot_creation: bool = False):
|
||||||
client = TLS(client_identifier='chrome110')
|
client = TLS(client_identifier='chrome110')
|
||||||
|
@ -24,40 +24,40 @@ class Account:
|
||||||
if logging: print('email', mail_address)
|
if logging: print('email', mail_address)
|
||||||
|
|
||||||
client.headers = {
|
client.headers = {
|
||||||
'authority' : 'poe.com',
|
'authority': 'poe.com',
|
||||||
'accept' : '*/*',
|
'accept': '*/*',
|
||||||
'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3',
|
'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3',
|
||||||
'content-type' : 'application/json',
|
'content-type': 'application/json',
|
||||||
'origin' : 'https://poe.com',
|
'origin': 'https://poe.com',
|
||||||
'poe-formkey' : 'null',
|
'poe-formkey': 'null',
|
||||||
'poe-tag-id' : 'null',
|
'poe-tag-id': 'null',
|
||||||
'poe-tchannel' : 'null',
|
'poe-tchannel': 'null',
|
||||||
'referer' : 'https://poe.com/login',
|
'referer': 'https://poe.com/login',
|
||||||
'sec-ch-ua' : '"Chromium";v="112", "Google Chrome";v="112", "Not:A-Brand";v="99"',
|
'sec-ch-ua': '"Chromium";v="112", "Google Chrome";v="112", "Not:A-Brand";v="99"',
|
||||||
'sec-ch-ua-mobile' : '?0',
|
'sec-ch-ua-mobile': '?0',
|
||||||
'sec-ch-ua-platform': '"macOS"',
|
'sec-ch-ua-platform': '"macOS"',
|
||||||
'sec-fetch-dest': 'empty',
|
'sec-fetch-dest': 'empty',
|
||||||
'sec-fetch-mode': 'cors',
|
'sec-fetch-mode': 'cors',
|
||||||
'sec-fetch-site': 'same-origin',
|
'sec-fetch-site': 'same-origin',
|
||||||
'user-agent' : 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36'
|
'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36'
|
||||||
}
|
}
|
||||||
|
|
||||||
client.headers["poe-formkey"] = extract_formkey(client.get('https://poe.com/login').text)
|
client.headers["poe-formkey"] = extract_formkey(client.get('https://poe.com/login').text)
|
||||||
client.headers["poe-tchannel"] = client.get('https://poe.com/api/settings').json()['tchannelData']['channel']
|
client.headers["poe-tchannel"] = client.get('https://poe.com/api/settings').json()['tchannelData']['channel']
|
||||||
|
|
||||||
#token = reCaptchaV3('https://www.recaptcha.net/recaptcha/enterprise/anchor?ar=1&k=6LflhEElAAAAAI_ewVwRWI9hsyV4mbZnYAslSvlG&co=aHR0cHM6Ly9wb2UuY29tOjQ0Mw..&hl=en&v=4PnKmGB9wRHh1i04o7YUICeI&size=invisible&cb=bi6ivxoskyal')
|
# token = reCaptchaV3('https://www.recaptcha.net/recaptcha/enterprise/anchor?ar=1&k=6LflhEElAAAAAI_ewVwRWI9hsyV4mbZnYAslSvlG&co=aHR0cHM6Ly9wb2UuY29tOjQ0Mw..&hl=en&v=4PnKmGB9wRHh1i04o7YUICeI&size=invisible&cb=bi6ivxoskyal')
|
||||||
token = solver.recaptcha(sitekey='6LflhEElAAAAAI_ewVwRWI9hsyV4mbZnYAslSvlG',
|
token = solver.recaptcha(sitekey='6LflhEElAAAAAI_ewVwRWI9hsyV4mbZnYAslSvlG',
|
||||||
url = 'https://poe.com/login?redirect_url=%2F',
|
url='https://poe.com/login?redirect_url=%2F',
|
||||||
version = 'v3',
|
version='v3',
|
||||||
enterprise = 1,
|
enterprise=1,
|
||||||
invisible = 1,
|
invisible=1,
|
||||||
action = 'login',)['code']
|
action='login', )['code']
|
||||||
|
|
||||||
payload = dumps(separators = (',', ':'), obj = {
|
payload = dumps(separators=(',', ':'), obj={
|
||||||
'queryName': 'MainSignupLoginSection_sendVerificationCodeMutation_Mutation',
|
'queryName': 'MainSignupLoginSection_sendVerificationCodeMutation_Mutation',
|
||||||
'variables': {
|
'variables': {
|
||||||
'emailAddress' : mail_address,
|
'emailAddress': mail_address,
|
||||||
'phoneNumber' : None,
|
'phoneNumber': None,
|
||||||
'recaptchaToken': token
|
'recaptchaToken': token
|
||||||
},
|
},
|
||||||
'query': 'mutation MainSignupLoginSection_sendVerificationCodeMutation_Mutation(\n $emailAddress: String\n $phoneNumber: String\n $recaptchaToken: String\n) {\n sendVerificationCode(verificationReason: login, emailAddress: $emailAddress, phoneNumber: $phoneNumber, recaptchaToken: $recaptchaToken) {\n status\n errorMessage\n }\n}\n',
|
'query': 'mutation MainSignupLoginSection_sendVerificationCodeMutation_Mutation(\n $emailAddress: String\n $phoneNumber: String\n $recaptchaToken: String\n) {\n sendVerificationCode(verificationReason: login, emailAddress: $emailAddress, phoneNumber: $phoneNumber, recaptchaToken: $recaptchaToken) {\n status\n errorMessage\n }\n}\n',
|
||||||
|
@ -74,17 +74,17 @@ class Account:
|
||||||
print('please try using a proxy / wait for fix')
|
print('please try using a proxy / wait for fix')
|
||||||
|
|
||||||
if 'Bad Request' in response.text:
|
if 'Bad Request' in response.text:
|
||||||
if logging: print('bad request, retrying...' , response.json())
|
if logging: print('bad request, retrying...', response.json())
|
||||||
quit()
|
quit()
|
||||||
|
|
||||||
if logging: print('send_code' ,response.json())
|
if logging: print('send_code', response.json())
|
||||||
|
|
||||||
mail_content = mail_client.get_message()
|
mail_content = mail_client.get_message()
|
||||||
mail_token = findall(r';">(\d{6,7})</div>', mail_content)[0]
|
mail_token = findall(r';">(\d{6,7})</div>', mail_content)[0]
|
||||||
|
|
||||||
if logging: print('code', mail_token)
|
if logging: print('code', mail_token)
|
||||||
|
|
||||||
payload = dumps(separators = (',', ':'), obj={
|
payload = dumps(separators=(',', ':'), obj={
|
||||||
"queryName": "SignupOrLoginWithCodeSection_signupWithVerificationCodeMutation_Mutation",
|
"queryName": "SignupOrLoginWithCodeSection_signupWithVerificationCodeMutation_Mutation",
|
||||||
"variables": {
|
"variables": {
|
||||||
"verificationCode": str(mail_token),
|
"verificationCode": str(mail_token),
|
||||||
|
@ -97,8 +97,8 @@ class Account:
|
||||||
base_string = payload + client.headers["poe-formkey"] + 'WpuLMiXEKKE98j56k'
|
base_string = payload + client.headers["poe-formkey"] + 'WpuLMiXEKKE98j56k'
|
||||||
client.headers["poe-tag-id"] = md5(base_string.encode()).hexdigest()
|
client.headers["poe-tag-id"] = md5(base_string.encode()).hexdigest()
|
||||||
|
|
||||||
response = client.post('https://poe.com/api/gql_POST', data = payload)
|
response = client.post('https://poe.com/api/gql_POST', data=payload)
|
||||||
if logging: print('verify_code', response.json())
|
if logging: print('verify_code', response.json())
|
||||||
|
|
||||||
|
|
||||||
Account.create(proxy = 'xtekky:wegwgwegwed_streaming-1@geo.iproyal.com:12321', logging = True)
|
Account.create(proxy='xtekky:wegwgwegwed_streaming-1@geo.iproyal.com:12321', logging=True)
|
||||||
|
|
|
@ -1,13 +1,13 @@
|
||||||
import quora
|
|
||||||
from time import sleep
|
from time import sleep
|
||||||
|
|
||||||
token = quora.Account.create(proxy = None,logging = True)
|
import quora
|
||||||
|
|
||||||
|
token = quora.Account.create(proxy=None, logging=True)
|
||||||
print('token', token)
|
print('token', token)
|
||||||
|
|
||||||
sleep(2)
|
sleep(2)
|
||||||
|
|
||||||
for response in quora.StreamingCompletion.create(model = 'gpt-3.5-turbo',
|
for response in quora.StreamingCompletion.create(model='gpt-3.5-turbo',
|
||||||
prompt = 'hello world',
|
prompt='hello world',
|
||||||
token = token):
|
token=token):
|
||||||
|
|
||||||
print(response.completion.choices[0].text, end="", flush=True)
|
print(response.completion.choices[0].text, end="", flush=True)
|
|
@ -1,18 +1,17 @@
|
||||||
import quora
|
import quora
|
||||||
|
|
||||||
token = quora.Account.create(logging = True, enable_bot_creation=True)
|
token = quora.Account.create(logging=True, enable_bot_creation=True)
|
||||||
|
|
||||||
model = quora.Model.create(
|
model = quora.Model.create(
|
||||||
token = token,
|
token=token,
|
||||||
model = 'gpt-3.5-turbo', # or claude-instant-v1.0
|
model='gpt-3.5-turbo', # or claude-instant-v1.0
|
||||||
system_prompt = 'you are ChatGPT a large language model ...'
|
system_prompt='you are ChatGPT a large language model ...'
|
||||||
)
|
)
|
||||||
|
|
||||||
print(model.name)
|
print(model.name)
|
||||||
|
|
||||||
for response in quora.StreamingCompletion.create(
|
for response in quora.StreamingCompletion.create(
|
||||||
custom_model = model.name,
|
custom_model=model.name,
|
||||||
prompt ='hello world',
|
prompt='hello world',
|
||||||
token = token):
|
token=token):
|
||||||
|
|
||||||
print(response.completion.choices[0].text)
|
print(response.completion.choices[0].text)
|
|
@ -1,7 +1,6 @@
|
||||||
import sqlchat
|
import sqlchat
|
||||||
|
|
||||||
for response in sqlchat.StreamCompletion.create(
|
for response in sqlchat.StreamCompletion.create(
|
||||||
prompt = 'write python code to reverse a string',
|
prompt='write python code to reverse a string',
|
||||||
messages = []):
|
messages=[]):
|
||||||
|
|
||||||
print(response.completion.choices[0].text, end='')
|
print(response.completion.choices[0].text, end='')
|
|
@ -1,7 +1,6 @@
|
||||||
import t3nsor
|
import t3nsor
|
||||||
|
|
||||||
for response in t3nsor.StreamCompletion.create(
|
for response in t3nsor.StreamCompletion.create(
|
||||||
prompt = 'write python code to reverse a string',
|
prompt='write python code to reverse a string',
|
||||||
messages = []):
|
messages=[]):
|
||||||
|
|
||||||
print(response.completion.choices[0].text)
|
print(response.completion.choices[0].text)
|
||||||
|
|
|
@ -2,18 +2,18 @@
|
||||||
import writesonic
|
import writesonic
|
||||||
|
|
||||||
# create account (3-4s)
|
# create account (3-4s)
|
||||||
account = writesonic.Account.create(logging = True)
|
account = writesonic.Account.create(logging=True)
|
||||||
|
|
||||||
# with loging:
|
# with loging:
|
||||||
# 2023-04-06 21:50:25 INFO __main__ -> register success : '{"id":"51aa0809-3053-44f7-922a...' (2s)
|
# 2023-04-06 21:50:25 INFO __main__ -> register success : '{"id":"51aa0809-3053-44f7-922a...' (2s)
|
||||||
# 2023-04-06 21:50:25 INFO __main__ -> id : '51aa0809-3053-44f7-922a-2b85d8d07edf'
|
# 2023-04-06 21:50:25 INFO __main__ -> id : '51aa0809-3053-44f7-922a-2b85d8d07edf'
|
||||||
# 2023-04-06 21:50:25 INFO __main__ -> token : 'eyJhbGciOiJIUzI1NiIsInR5cCI6Ik...'
|
# 2023-04-06 21:50:25 INFO __main__ -> token : 'eyJhbGciOiJIUzI1NiIsInR5cCI6Ik...'
|
||||||
# 2023-04-06 21:50:28 INFO __main__ -> got key : '194158c4-d249-4be0-82c6-5049e869533c' (2s)
|
# 2023-04-06 21:50:28 INFO __main__ -> got key : '194158c4-d249-4be0-82c6-5049e869533c' (2s)
|
||||||
|
|
||||||
# simple completion
|
# simple completion
|
||||||
response = writesonic.Completion.create(
|
response = writesonic.Completion.create(
|
||||||
api_key = account.key,
|
api_key=account.key,
|
||||||
prompt = 'hello world'
|
prompt='hello world'
|
||||||
)
|
)
|
||||||
|
|
||||||
print(response.completion.choices[0].text) # Hello! How may I assist you today?
|
print(response.completion.choices[0].text) # Hello! How may I assist you today?
|
||||||
|
@ -21,10 +21,10 @@ print(response.completion.choices[0].text) # Hello! How may I assist you today?
|
||||||
# conversation
|
# conversation
|
||||||
|
|
||||||
response = writesonic.Completion.create(
|
response = writesonic.Completion.create(
|
||||||
api_key = account.key,
|
api_key=account.key,
|
||||||
prompt = 'what is my name ?',
|
prompt='what is my name ?',
|
||||||
enable_memory = True,
|
enable_memory=True,
|
||||||
history_data = [
|
history_data=[
|
||||||
{
|
{
|
||||||
'is_sent': True,
|
'is_sent': True,
|
||||||
'message': 'my name is Tekky'
|
'message': 'my name is Tekky'
|
||||||
|
@ -41,9 +41,9 @@ print(response.completion.choices[0].text) # Your name is Tekky.
|
||||||
# enable internet
|
# enable internet
|
||||||
|
|
||||||
response = writesonic.Completion.create(
|
response = writesonic.Completion.create(
|
||||||
api_key = account.key,
|
api_key=account.key,
|
||||||
prompt = 'who won the quatar world cup ?',
|
prompt='who won the quatar world cup ?',
|
||||||
enable_google_results = True
|
enable_google_results=True
|
||||||
)
|
)
|
||||||
|
|
||||||
print(response.completion.choices[0].text) # Argentina won the 2022 FIFA World Cup tournament held in Qatar ...
|
print(response.completion.choices[0].text) # Argentina won the 2022 FIFA World Cup tournament held in Qatar ...
|
|
@ -1,12 +1,12 @@
|
||||||
from requests import Session
|
|
||||||
from re import search
|
|
||||||
from random import randint
|
|
||||||
from json import dumps, loads
|
from json import dumps, loads
|
||||||
from urllib.parse import urlencode
|
|
||||||
from dotenv import load_dotenv
|
|
||||||
from os import getenv
|
from os import getenv
|
||||||
|
from random import randint
|
||||||
|
from re import search
|
||||||
|
from urllib.parse import urlencode
|
||||||
|
|
||||||
from bard.typings import BardResponse
|
from bard.typings import BardResponse
|
||||||
|
from dotenv import load_dotenv
|
||||||
|
from requests import Session
|
||||||
|
|
||||||
load_dotenv()
|
load_dotenv()
|
||||||
token = getenv('1psid')
|
token = getenv('1psid')
|
||||||
|
@ -62,7 +62,8 @@ class Completion:
|
||||||
'rt': 'c',
|
'rt': 'c',
|
||||||
})
|
})
|
||||||
|
|
||||||
response = client.post(f'https://bard.google.com/_/BardChatUi/data/assistant.lamda.BardFrontendService/StreamGenerate?{params}',
|
response = client.post(
|
||||||
|
f'https://bard.google.com/_/BardChatUi/data/assistant.lamda.BardFrontendService/StreamGenerate?{params}',
|
||||||
data={
|
data={
|
||||||
'at': snlm0e,
|
'at': snlm0e,
|
||||||
'f.req': dumps([None, dumps([
|
'f.req': dumps([None, dumps([
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
from typing import Dict, List, Optional, Union
|
from typing import Dict, List, Union
|
||||||
|
|
||||||
|
|
||||||
class BardResponse:
|
class BardResponse:
|
||||||
|
|
|
@ -1,14 +1,12 @@
|
||||||
# Import necessary libraries
|
# Import necessary libraries
|
||||||
from requests import get
|
|
||||||
from browser_cookie3 import edge, chrome
|
|
||||||
from ssl import create_default_context
|
|
||||||
from certifi import where
|
|
||||||
from uuid import uuid4
|
|
||||||
from random import randint
|
|
||||||
from json import dumps, loads
|
|
||||||
|
|
||||||
import asyncio
|
import asyncio
|
||||||
|
from json import dumps, loads
|
||||||
|
from ssl import create_default_context
|
||||||
|
|
||||||
import websockets
|
import websockets
|
||||||
|
from browser_cookie3 import edge
|
||||||
|
from certifi import where
|
||||||
|
from requests import get
|
||||||
|
|
||||||
# Set up SSL context
|
# Set up SSL context
|
||||||
ssl_context = create_default_context()
|
ssl_context = create_default_context()
|
||||||
|
@ -83,7 +81,7 @@ class AsyncCompletion:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
response = loads(obj)
|
response = loads(obj)
|
||||||
if response.get('type') == 1 and response['arguments'][0].get('messages',):
|
if response.get('type') == 1 and response['arguments'][0].get('messages', ):
|
||||||
response_text = response['arguments'][0]['messages'][0]['adaptiveCards'][0]['body'][0].get(
|
response_text = response['arguments'][0]['messages'][0]['adaptiveCards'][0]['body'][0].get(
|
||||||
'text')
|
'text')
|
||||||
|
|
||||||
|
@ -106,4 +104,5 @@ async def run():
|
||||||
):
|
):
|
||||||
print(value, end='', flush=True)
|
print(value, end='', flush=True)
|
||||||
|
|
||||||
|
|
||||||
asyncio.run(run())
|
asyncio.run(run())
|
||||||
|
|
|
@ -6,7 +6,6 @@ class Completion:
|
||||||
system_prompt=("ASSUME I HAVE FULL ACCESS TO COCALC. ENCLOSE MATH IN $. "
|
system_prompt=("ASSUME I HAVE FULL ACCESS TO COCALC. ENCLOSE MATH IN $. "
|
||||||
"INCLUDE THE LANGUAGE DIRECTLY AFTER THE TRIPLE BACKTICKS "
|
"INCLUDE THE LANGUAGE DIRECTLY AFTER THE TRIPLE BACKTICKS "
|
||||||
"IN ALL MARKDOWN CODE BLOCKS. How can I do the following using CoCalc?")) -> str:
|
"IN ALL MARKDOWN CODE BLOCKS. How can I do the following using CoCalc?")) -> str:
|
||||||
|
|
||||||
# Initialize a session with custom headers
|
# Initialize a session with custom headers
|
||||||
session = self._initialize_session()
|
session = self._initialize_session()
|
||||||
|
|
||||||
|
|
|
@ -1,8 +1,7 @@
|
||||||
import cocalc
|
import cocalc
|
||||||
|
|
||||||
|
|
||||||
response = cocalc.Completion.create(
|
response = cocalc.Completion.create(
|
||||||
prompt = 'hello world'
|
prompt='hello world'
|
||||||
)
|
)
|
||||||
|
|
||||||
print(response)
|
print(response)
|
|
@ -1,7 +1,8 @@
|
||||||
# Import necessary libraries
|
# Import necessary libraries
|
||||||
from requests import get
|
|
||||||
from os import urandom
|
|
||||||
from json import loads
|
from json import loads
|
||||||
|
from os import urandom
|
||||||
|
|
||||||
|
from requests import get
|
||||||
|
|
||||||
# Generate a random session ID
|
# Generate a random session ID
|
||||||
sessionId = urandom(10).hex()
|
sessionId = urandom(10).hex()
|
||||||
|
|
|
@ -1,6 +1,8 @@
|
||||||
import websockets
|
|
||||||
from json import dumps, loads
|
from json import dumps, loads
|
||||||
|
|
||||||
|
import websockets
|
||||||
|
|
||||||
|
|
||||||
# Define the asynchronous function to test the WebSocket connection
|
# Define the asynchronous function to test the WebSocket connection
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,7 +1,8 @@
|
||||||
# Import required libraries
|
# Import required libraries
|
||||||
from tls_client import Session
|
|
||||||
from uuid import uuid4
|
from uuid import uuid4
|
||||||
|
|
||||||
from browser_cookie3 import chrome
|
from browser_cookie3 import chrome
|
||||||
|
from tls_client import Session
|
||||||
|
|
||||||
|
|
||||||
class OpenAIChat:
|
class OpenAIChat:
|
||||||
|
|
|
@ -1,7 +1,8 @@
|
||||||
import requests
|
|
||||||
import json
|
import json
|
||||||
import re
|
import re
|
||||||
|
|
||||||
|
import requests
|
||||||
|
|
||||||
headers = {
|
headers = {
|
||||||
'authority': 'openai.a2hosted.com',
|
'authority': 'openai.a2hosted.com',
|
||||||
'accept': 'text/event-stream',
|
'accept': 'text/event-stream',
|
||||||
|
@ -13,10 +14,12 @@ headers = {
|
||||||
'user-agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 Edg/113.0.0.0',
|
'user-agent': 'Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/113.0.0.0 Safari/537.36 Edg/113.0.0.0',
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
def create_query_param(conversation):
|
def create_query_param(conversation):
|
||||||
encoded_conversation = json.dumps(conversation)
|
encoded_conversation = json.dumps(conversation)
|
||||||
return encoded_conversation.replace(" ", "%20").replace('"', '%22').replace("'", "%27")
|
return encoded_conversation.replace(" ", "%20").replace('"', '%22').replace("'", "%27")
|
||||||
|
|
||||||
|
|
||||||
user_input = input("Enter your message: ")
|
user_input = input("Enter your message: ")
|
||||||
|
|
||||||
data = [
|
data = [
|
||||||
|
|
|
@ -1,9 +1,9 @@
|
||||||
from requests import post, get
|
|
||||||
from json import dumps
|
from json import dumps
|
||||||
#from mail import MailClient
|
# from mail import MailClient
|
||||||
from time import sleep
|
|
||||||
from re import findall
|
from re import findall
|
||||||
|
|
||||||
|
from requests import post, get
|
||||||
|
|
||||||
html = get('https://developermail.com/mail/')
|
html = get('https://developermail.com/mail/')
|
||||||
print(html.cookies.get('mailboxId'))
|
print(html.cookies.get('mailboxId'))
|
||||||
email = findall(r'mailto:(.*)">', html.text)[0]
|
email = findall(r'mailto:(.*)">', html.text)[0]
|
||||||
|
@ -15,9 +15,9 @@ headers = {
|
||||||
}
|
}
|
||||||
|
|
||||||
json_data = {
|
json_data = {
|
||||||
'email' : email,
|
'email': email,
|
||||||
'password': 'T4xyt4Yn6WWQ4NC',
|
'password': 'T4xyt4Yn6WWQ4NC',
|
||||||
'data' : {},
|
'data': {},
|
||||||
'gotrue_meta_security': {},
|
'gotrue_meta_security': {},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1,6 +1,8 @@
|
||||||
import requests
|
|
||||||
import email
|
import email
|
||||||
|
|
||||||
|
import requests
|
||||||
|
|
||||||
|
|
||||||
class MailClient:
|
class MailClient:
|
||||||
|
|
||||||
def __init__(self):
|
def __init__(self):
|
||||||
|
|
|
@ -30,8 +30,7 @@ json_data = {
|
||||||
],
|
],
|
||||||
}
|
}
|
||||||
|
|
||||||
response = requests.post('https://openprompt.co/api/chat2', cookies=cookies, headers=headers, json=json_data, stream=True)
|
response = requests.post('https://openprompt.co/api/chat2', cookies=cookies, headers=headers, json=json_data,
|
||||||
|
stream=True)
|
||||||
for chunk in response.iter_content(chunk_size=1024):
|
for chunk in response.iter_content(chunk_size=1024):
|
||||||
print(chunk)
|
print(chunk)
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -1,7 +1,6 @@
|
||||||
access_token = 'eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV'
|
access_token = 'eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV'
|
||||||
supabase_auth_token= '%5B%22eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV8%22%2C%22_Zp8uXIA2InTDKYgo8TCqA%22%2Cnull%2Cnull%2Cnull%5D'
|
supabase_auth_token = '%5B%22eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV8%22%2C%22_Zp8uXIA2InTDKYgo8TCqA%22%2Cnull%2Cnull%2Cnull%5D'
|
||||||
|
|
||||||
|
|
||||||
idk = [
|
idk = [
|
||||||
"eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV8",
|
"eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJhdWQiOiJhdXRoZW50aWNhdGVkIiwiZXhwIjoxNjgyMjk0ODcxLCJzdWIiOiI4NWNkNTNiNC1lZTUwLTRiMDQtOGJhNS0wNTUyNjk4ODliZDIiLCJlbWFpbCI6ImNsc2J5emdqcGhiQGJ1Z2Zvby5jb20iLCJwaG9uZSI6IiIsImFwcF9tZXRhZGF0YSI6eyJwcm92aWRlciI6ImVtYWlsIiwicHJvdmlkZXJzIjpbImVtYWlsIl19LCJ1c2VyX21ldGFkYXRhIjp7fSwicm9sZSI6ImF1dGhlbnRpY2F0ZWQiLCJhYWwiOiJhYWwxIiwiYW1yIjpbeyJtZXRob2QiOiJvdHAiLCJ0aW1lc3RhbXAiOjE2ODE2OTAwNzF9XSwic2Vzc2lvbl9pZCI6ImY4MTg1YTM5LTkxYzgtNGFmMy1iNzAxLTdhY2MwY2MwMGNlNSJ9.UvcTfpyIM1TdzM8ZV6UAPWfa0rgNq4AiqeD0INy6zV8",
|
||||||
"_Zp8uXIA2InTDKYgo8TCqA",None,None,None]
|
"_Zp8uXIA2InTDKYgo8TCqA", None, None, None]
|
||||||
|
|
|
@ -1,6 +1,7 @@
|
||||||
from requests import post
|
|
||||||
from time import time
|
from time import time
|
||||||
|
|
||||||
|
from requests import post
|
||||||
|
|
||||||
headers = {
|
headers = {
|
||||||
'authority': 'www.t3nsor.tech',
|
'authority': 'www.t3nsor.tech',
|
||||||
'accept': '*/*',
|
'accept': '*/*',
|
||||||
|
@ -19,10 +20,9 @@ headers = {
|
||||||
'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36',
|
'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36',
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
class T3nsorResponse:
|
class T3nsorResponse:
|
||||||
|
|
||||||
class Completion:
|
class Completion:
|
||||||
|
|
||||||
class Choices:
|
class Choices:
|
||||||
def __init__(self, choice: dict) -> None:
|
def __init__(self, choice: dict) -> None:
|
||||||
self.text = choice['text']
|
self.text = choice['text']
|
||||||
|
@ -47,7 +47,6 @@ class T3nsorResponse:
|
||||||
return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>'''
|
return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>'''
|
||||||
|
|
||||||
def __init__(self, response_dict: dict) -> None:
|
def __init__(self, response_dict: dict) -> None:
|
||||||
|
|
||||||
self.response_dict = response_dict
|
self.response_dict = response_dict
|
||||||
self.id = response_dict['id']
|
self.id = response_dict['id']
|
||||||
self.object = response_dict['object']
|
self.object = response_dict['object']
|
||||||
|
@ -59,79 +58,79 @@ class T3nsorResponse:
|
||||||
def json(self) -> dict:
|
def json(self) -> dict:
|
||||||
return self.response_dict
|
return self.response_dict
|
||||||
|
|
||||||
|
|
||||||
class Completion:
|
class Completion:
|
||||||
model = {
|
model = {
|
||||||
'model': {
|
'model': {
|
||||||
'id' : 'gpt-3.5-turbo',
|
'id': 'gpt-3.5-turbo',
|
||||||
'name' : 'Default (GPT-3.5)'
|
'name': 'Default (GPT-3.5)'
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
def create(
|
def create(
|
||||||
prompt: str = 'hello world',
|
prompt: str = 'hello world',
|
||||||
messages: list = []) -> T3nsorResponse:
|
messages: list = []) -> T3nsorResponse:
|
||||||
|
response = post('https://www.t3nsor.tech/api/chat', headers=headers, json=Completion.model | {
|
||||||
response = post('https://www.t3nsor.tech/api/chat', headers = headers, json = Completion.model | {
|
'messages': messages,
|
||||||
'messages' : messages,
|
'key': '',
|
||||||
'key' : '',
|
'prompt': prompt
|
||||||
'prompt' : prompt
|
|
||||||
})
|
})
|
||||||
|
|
||||||
return T3nsorResponse({
|
return T3nsorResponse({
|
||||||
'id' : f'cmpl-1337-{int(time())}',
|
'id': f'cmpl-1337-{int(time())}',
|
||||||
'object' : 'text_completion',
|
'object': 'text_completion',
|
||||||
'created': int(time()),
|
'created': int(time()),
|
||||||
'model' : Completion.model,
|
'model': Completion.model,
|
||||||
'choices': [{
|
'choices': [{
|
||||||
'text' : response.text,
|
'text': response.text,
|
||||||
'index' : 0,
|
'index': 0,
|
||||||
'logprobs' : None,
|
'logprobs': None,
|
||||||
'finish_reason' : 'stop'
|
'finish_reason': 'stop'
|
||||||
}],
|
}],
|
||||||
'usage': {
|
'usage': {
|
||||||
'prompt_chars' : len(prompt),
|
'prompt_chars': len(prompt),
|
||||||
'completion_chars' : len(response.text),
|
'completion_chars': len(response.text),
|
||||||
'total_chars' : len(prompt) + len(response.text)
|
'total_chars': len(prompt) + len(response.text)
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
|
|
||||||
class StreamCompletion:
|
class StreamCompletion:
|
||||||
model = {
|
model = {
|
||||||
'model': {
|
'model': {
|
||||||
'id' : 'gpt-3.5-turbo',
|
'id': 'gpt-3.5-turbo',
|
||||||
'name' : 'Default (GPT-3.5)'
|
'name': 'Default (GPT-3.5)'
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
def create(
|
def create(
|
||||||
prompt: str = 'hello world',
|
prompt: str = 'hello world',
|
||||||
messages: list = []) -> T3nsorResponse:
|
messages: list = []) -> T3nsorResponse:
|
||||||
|
|
||||||
print('t3nsor api is down, this may not work, refer to another module')
|
print('t3nsor api is down, this may not work, refer to another module')
|
||||||
|
|
||||||
response = post('https://www.t3nsor.tech/api/chat', headers = headers, stream = True, json = Completion.model | {
|
response = post('https://www.t3nsor.tech/api/chat', headers=headers, stream=True, json=Completion.model | {
|
||||||
'messages' : messages,
|
'messages': messages,
|
||||||
'key' : '',
|
'key': '',
|
||||||
'prompt' : prompt
|
'prompt': prompt
|
||||||
})
|
})
|
||||||
|
|
||||||
for chunk in response.iter_content(chunk_size = 2046):
|
for chunk in response.iter_content(chunk_size=2046):
|
||||||
yield T3nsorResponse({
|
yield T3nsorResponse({
|
||||||
'id' : f'cmpl-1337-{int(time())}',
|
'id': f'cmpl-1337-{int(time())}',
|
||||||
'object' : 'text_completion',
|
'object': 'text_completion',
|
||||||
'created': int(time()),
|
'created': int(time()),
|
||||||
'model' : Completion.model,
|
'model': Completion.model,
|
||||||
|
|
||||||
'choices': [{
|
'choices': [{
|
||||||
'text' : chunk.decode(),
|
'text': chunk.decode(),
|
||||||
'index' : 0,
|
'index': 0,
|
||||||
'logprobs' : None,
|
'logprobs': None,
|
||||||
'finish_reason' : 'stop'
|
'finish_reason': 'stop'
|
||||||
}],
|
}],
|
||||||
|
|
||||||
'usage': {
|
'usage': {
|
||||||
'prompt_chars' : len(prompt),
|
'prompt_chars': len(prompt),
|
||||||
'completion_chars' : len(chunk.decode()),
|
'completion_chars': len(chunk.decode()),
|
||||||
'total_chars' : len(prompt) + len(chunk.decode())
|
'total_chars': len(prompt) + len(chunk.decode())
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
|
|
|
@ -1,12 +1,8 @@
|
||||||
import gptbz
|
|
||||||
import asyncio
|
|
||||||
|
|
||||||
|
|
||||||
# asyncio.run(gptbz.test())
|
# asyncio.run(gptbz.test())
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
|
|
||||||
image = '/9j/4AAQSkZJRgABAQEAYABgAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDAxNDQ0Hyc5PTgyPC4zNDL/2wBDAQkJCQwLDBgNDRgyIRwhMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjL/wAARCAAoALQDASIAAhEBAxEB/8QAHwAAAQUBAQEBAQEAAAAAAAAAAAECAwQFBgcICQoL/8QAtRAAAgEDAwIEAwUFBAQAAAF9AQIDAAQRBRIhMUEGE1FhByJxFDKBkaEII0KxwRVS0fAkM2JyggkKFhcYGRolJicoKSo0NTY3ODk6Q0RFRkdISUpTVFVWV1hZWmNkZWZnaGlqc3R1dnd4eXqDhIWGh4iJipKTlJWWl5iZmqKjpKWmp6ipqrKztLW2t7i5usLDxMXGx8jJytLT1NXW19jZ2uHi4+Tl5ufo6erx8vP09fb3+Pn6/8QAHwEAAwEBAQEBAQEBAQAAAAAAAAECAwQFBgcICQoL/8QAtREAAgECBAQDBAcFBAQAAQJ3AAECAxEEBSExBhJBUQdhcRMiMoEIFEKRobHBCSMzUvAVYnLRChYkNOEl8RcYGRomJygpKjU2Nzg5OkNERUZHSElKU1RVVldYWVpjZGVmZ2hpanN0dXZ3eHl6goOEhYaHiImKkpOUlZaXmJmaoqOkpaanqKmqsrO0tba3uLm6wsPExcbHyMnK0tPU1dbX2Nna4uPk5ebn6Onq8vP09fb3+Pn6/9oADAMBAAIRAxEAPwD3+iiigDkZP+EhS4W0k1S+VntQPtEWmRsgkNwBu4ZsHYQNvTbls5BA6DS7uW6S6E0VwjQ3UsQM0Pl71DZUrydy4IAbvg8CsTx3DbHQLi4uVs9scWzdd+dsAaWI4PlfNjKjpzkDtmpoNSgbWYpLR7Ty5bq5trw/vd3nIowBxtzti53Y6fKT3z2djra56fNbv07HR1z13ZRX/jDyby0+02f9nfdmsEeHd5o/5anndwPkxjjPWuhrh9Mvra88RLqccmnOHtvLEqfaN+1r1lUcjbg4PbO4H+Cqk+hnRi9ZI29E0uC2N1eG3Am+13DITZRwuqlsYG0ZYEKCGJywwT2AtWTapcW1vcPPCiyrE5ils2SRQV+dW/ecMT/3zgj5utZtpdwL4e190e02W9xeb9vm7FOWY78/NnnJ28f3ahkgtptD8JRlbMos9s8QPnbcrEzDy/4sgDjzOMdeaSZbi23f8vmbfn6hBFuktmuWWPJWCNELNuxgbpcDj1Pbr2qJ9bMVyIZNK1JVLyr5qwB1AjUNu+Uk4bovGSRjAqCTwdoElv5B02MReT5G1HZfk8zzMcEfx81YlsJ7NJX0tolZzNK8dyZJA8jDIwd3yjcBkAHjOAM09SP3b/q36mkjiSNXAYBgCNykH8QeRWdfaw1ldSW66XqN0UgE++3iBRsvt2BiQN/8WPQZqharF9oN5osVml1NLbLqUbmUFY/L4CrgYYKy4yoGM5xjhlnc2OoeMrfULV7aQXGkExyYlErJ5oPQ/Jtye/zZ9qLgqaTba0NyzvPtizH7NcQeVM8OJ49u/acbl9VPY96s1geFjF/xOhF9m41Wfd9n8z73BO7f/Fzzt+X0q7c6mWvRY2DwSXcUsQuUff8Auo2ySflB+YqrYyQOmTyARPQmVP32kLqF1cbmsrJZkuni3rcfZ98UfzKvJJUE4JOM5wpODwDl3Meuf2rHbRatcBJXuj5iachjhUovlBmZudrNkEZ3HIOMGlhREhbS9He2a8MO6a4fzmGDMQ3zAk5yZ8DzMgj0yRuWdha2CzLawrEJpnnkx/G7HLMfc0bl3VNf5pff/kVLS8uxFHHJZ3s5Xyo2mZI4y2VBZyN44B6gDrwAcVZ069Go2EV2Le5t/MBPlXMZjkXnGGU9OlULSdbfTt8LWy5mt0JAkK4YRLjnnODx26Z71TXULEWn/CUWDwmxeDbM4WbkCXJbaB23SnlM5PUDNF7CcObZf12OlpCcDoTz2oVlcZVgRkjIPccGo7hgsSk7ceYg+bP94elUYpamda64915GdH1SESxiTM0KjZmTZtbDHB53Y/u89eK1qw4xD9l0mIC3wLdCg/eYwHh+73x0+9znb71uUkXUSWyCiiimZhRRRQBieL5Hj8LXjxySxuNmGivFtWHzr0lbhfx69O9MvHdZpbKKWYnUluNji+VGikVFULHnkdGbjO05JHPEviyF5/DF7HGkjuQpCx2i3THDA8RNw3Tv069qR0kk0i4uFilF3bSXTwE2a+YGzIAUQnnIPByN46kbjUPc6YNKC9X+SLtjeB9Mt5ZyqzbI1lQzK5R2C/KWGAT8w6dcjHUVzemSyxeCba9e5uWfzIgxl1aOTgXPebGw5BwR3ACdalna8+0R3Kx3nk6jc2MvkjTI2MH97zDnI+4uWOSny4z2Lqxmt/hytvHHIZhFHJsj0yJnyXDEfZ87M9cjPB56ik2y4xSsu7XcnjMsejeJszXBZZrgozaihZAYwQFfGIQM8Bvu9ehrTKuJtOg3y5gKs/8ApAy2Y5B846uMj8Tz/CaqzROH1C3EchW6uHGRZIVx9nHXs4yPvN1PydBV2Lc+u3eUkCJBDtZoAFJzJna/VjgjI/h/4EaaM5PS/wDXRF+iiirOcy7RZE8RanukmKPFA6q9yHVfvg7Y+qfd5J4Y9OhrJ8Nm4FxYJNNdORaXCsJtTS4yVnAyQoG5sfxfw/dPJrUslmGt6rcymQxM0MMStahMALk4cfM65c9cBSGA7mqmi2k9t/ZZuDJJKbSdpHNjHEdzyRvhtv3G5PyjIbBJOVqDpurP5d+zGWtzeLdahZQLNK895PiV7+N/IURKQQMEqNzKAm1tucnggG4Fkhs4INNuJL145oEuHa7BcIAuWOQRkrhiAFzkkEE8rNDJPczWtnG1rG7yfapvsqESsY1AIJPP3hztbPllTjHKvpv2CWKbTUSHdJCk8cVtH+8jUFOSNpGAynOTgJgL1BNRNxf9fmWNGa3fR7U2ty9zDswJZJxMzHvlwSCc5BwccVerBZ3tLf8Atqyguvsxt/n02OyUSsxk3FsHa24bnyM4ycgE9d1WDDIz1I5BHQ471SM6i1uY8cjjSIWLyFjLbDJu1J5Mefn6HryP4snH3hRdmTS5f7T82aS2WBY5Y5LpVjX94Pn+YYzhmydw4UDB4wio/wDY8K+XLuE1qcfY1B4MWfk6DHOT/Bg4+6K1zGkkHlSoroy7WVlGCCOQRSsU5JGUrPo96EZ5p7O7mmmlubm7XFqQoYIobB2fK3Aztwe3TQvX2QKQSMyxDiQJ1dR1P8u/TvWb5bWty2m3KTXlvqMs7Ky2ieVbqVBKSEcHJL4JB3ZwfeLfcQRnTpY7mT7PLZiOdbJSkillzgA44KMScLsBBAOBkuNxu0/6epcQv9s0+LfJzauxBuVJJDRckdXPJ+YcDJH8QrTrN2sNcsxsk2LZyjd9nXaCWj439VPH3RwcZ/hFaVNGc+gUUUUyAooooAxfFVxZxeG9RS7ltVQ25ytwzbCCQBkJ82MkD5eeah0G7tYLi/sZJrKO4fUbjy4oncM/SQ5D9Ww4J25Xniiis2/eO2FNOhf1/CxmamsEGp2+nzx2CwxajYyWKN9o3KdpX+Ebd2I2287ePm973i3UdMg0W+0y4mtUkNqJPKuBJ5ewuEBYx8gbiBxz+FFFS3ZM1p01OdNN/wBaFfVtU0qHxHplx9qsSkEl2853SvIjxwjdtCZXIX7wbt05q7YJdS6nc6vYxWEtpfi2KS+bKsjQhCSWBBG4bhtAAyCcmiinF3k0RWgqdKMl1VvxZfM2s+VkWFh5nl5x9tfG/djGfK6bec468Y/irN1CeUCeHXbrTItPc3O6GN5PNltxHx0I+YKXLYB42455ooqpaIwo2lO1rE1rZjUYrcCO2Giw/Zp7BYzKrkKu4bh8oAB2EA56HIz0u3uxL+1kbygQpQFt2fmki4GOOuOvfHbNFFPpcTu6nKFpsTU75V8oNJKXIXduOI4hk54zjHTjGO+a0KKKaM59PQxLqNNBMuoQpDFYJEfPQLISp8zcWAXIxh5CcLnOMnHQaFNKkkvtOFoli0k9xqP32Zn24LIFyM7kwRg98c5yUVL3No6xTfV2/IrxyW0vh21kQ2phaexKn97s5aErj+LPTbnj7u7+KujoopxZNZW+9/oQXdpBfWk1rcxiSGVGjdSSMhgQeRyOCRxWOtvbXU0Ol6mIHksJbea0IMoJYISGy3U5ST+JuB83uUUMVJuz121JnaL/AITOBSYPOGnyEA7/ADdvmJnH8G3IHX5s4xxmtmiihdRVFZR9AoooqjI//9k='
|
image = '/9j/4AAQSkZJRgABAQEAYABgAAD/2wBDAAgGBgcGBQgHBwcJCQgKDBQNDAsLDBkSEw8UHRofHh0aHBwgJC4nICIsIxwcKDcpLDAxNDQ0Hyc5PTgyPC4zNDL/2wBDAQkJCQwLDBgNDRgyIRwhMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjIyMjL/wAARCAAoALQDASIAAhEBAxEB/8QAHwAAAQUBAQEBAQEAAAAAAAAAAAECAwQFBgcICQoL/8QAtRAAAgEDAwIEAwUFBAQAAAF9AQIDAAQRBRIhMUEGE1FhByJxFDKBkaEII0KxwRVS0fAkM2JyggkKFhcYGRolJicoKSo0NTY3ODk6Q0RFRkdISUpTVFVWV1hZWmNkZWZnaGlqc3R1dnd4eXqDhIWGh4iJipKTlJWWl5iZmqKjpKWmp6ipqrKztLW2t7i5usLDxMXGx8jJytLT1NXW19jZ2uHi4+Tl5ufo6erx8vP09fb3+Pn6/8QAHwEAAwEBAQEBAQEBAQAAAAAAAAECAwQFBgcICQoL/8QAtREAAgECBAQDBAcFBAQAAQJ3AAECAxEEBSExBhJBUQdhcRMiMoEIFEKRobHBCSMzUvAVYnLRChYkNOEl8RcYGRomJygpKjU2Nzg5OkNERUZHSElKU1RVVldYWVpjZGVmZ2hpanN0dXZ3eHl6goOEhYaHiImKkpOUlZaXmJmaoqOkpaanqKmqsrO0tba3uLm6wsPExcbHyMnK0tPU1dbX2Nna4uPk5ebn6Onq8vP09fb3+Pn6/9oADAMBAAIRAxEAPwD3+iiigDkZP+EhS4W0k1S+VntQPtEWmRsgkNwBu4ZsHYQNvTbls5BA6DS7uW6S6E0VwjQ3UsQM0Pl71DZUrydy4IAbvg8CsTx3DbHQLi4uVs9scWzdd+dsAaWI4PlfNjKjpzkDtmpoNSgbWYpLR7Ty5bq5trw/vd3nIowBxtzti53Y6fKT3z2djra56fNbv07HR1z13ZRX/jDyby0+02f9nfdmsEeHd5o/5anndwPkxjjPWuhrh9Mvra88RLqccmnOHtvLEqfaN+1r1lUcjbg4PbO4H+Cqk+hnRi9ZI29E0uC2N1eG3Am+13DITZRwuqlsYG0ZYEKCGJywwT2AtWTapcW1vcPPCiyrE5ils2SRQV+dW/ecMT/3zgj5utZtpdwL4e190e02W9xeb9vm7FOWY78/NnnJ28f3ahkgtptD8JRlbMos9s8QPnbcrEzDy/4sgDjzOMdeaSZbi23f8vmbfn6hBFuktmuWWPJWCNELNuxgbpcDj1Pbr2qJ9bMVyIZNK1JVLyr5qwB1AjUNu+Uk4bovGSRjAqCTwdoElv5B02MReT5G1HZfk8zzMcEfx81YlsJ7NJX0tolZzNK8dyZJA8jDIwd3yjcBkAHjOAM09SP3b/q36mkjiSNXAYBgCNykH8QeRWdfaw1ldSW66XqN0UgE++3iBRsvt2BiQN/8WPQZqharF9oN5osVml1NLbLqUbmUFY/L4CrgYYKy4yoGM5xjhlnc2OoeMrfULV7aQXGkExyYlErJ5oPQ/Jtye/zZ9qLgqaTba0NyzvPtizH7NcQeVM8OJ49u/acbl9VPY96s1geFjF/xOhF9m41Wfd9n8z73BO7f/Fzzt+X0q7c6mWvRY2DwSXcUsQuUff8Auo2ySflB+YqrYyQOmTyARPQmVP32kLqF1cbmsrJZkuni3rcfZ98UfzKvJJUE4JOM5wpODwDl3Meuf2rHbRatcBJXuj5iachjhUovlBmZudrNkEZ3HIOMGlhREhbS9He2a8MO6a4fzmGDMQ3zAk5yZ8DzMgj0yRuWdha2CzLawrEJpnnkx/G7HLMfc0bl3VNf5pff/kVLS8uxFHHJZ3s5Xyo2mZI4y2VBZyN44B6gDrwAcVZ069Go2EV2Le5t/MBPlXMZjkXnGGU9OlULSdbfTt8LWy5mt0JAkK4YRLjnnODx26Z71TXULEWn/CUWDwmxeDbM4WbkCXJbaB23SnlM5PUDNF7CcObZf12OlpCcDoTz2oVlcZVgRkjIPccGo7hgsSk7ceYg+bP94elUYpamda64915GdH1SESxiTM0KjZmTZtbDHB53Y/u89eK1qw4xD9l0mIC3wLdCg/eYwHh+73x0+9znb71uUkXUSWyCiiimZhRRRQBieL5Hj8LXjxySxuNmGivFtWHzr0lbhfx69O9MvHdZpbKKWYnUluNji+VGikVFULHnkdGbjO05JHPEviyF5/DF7HGkjuQpCx2i3THDA8RNw3Tv069qR0kk0i4uFilF3bSXTwE2a+YGzIAUQnnIPByN46kbjUPc6YNKC9X+SLtjeB9Mt5ZyqzbI1lQzK5R2C/KWGAT8w6dcjHUVzemSyxeCba9e5uWfzIgxl1aOTgXPebGw5BwR3ACdalna8+0R3Kx3nk6jc2MvkjTI2MH97zDnI+4uWOSny4z2Lqxmt/hytvHHIZhFHJsj0yJnyXDEfZ87M9cjPB56ik2y4xSsu7XcnjMsejeJszXBZZrgozaihZAYwQFfGIQM8Bvu9ehrTKuJtOg3y5gKs/8ApAy2Y5B846uMj8Tz/CaqzROH1C3EchW6uHGRZIVx9nHXs4yPvN1PydBV2Lc+u3eUkCJBDtZoAFJzJna/VjgjI/h/4EaaM5PS/wDXRF+iiirOcy7RZE8RanukmKPFA6q9yHVfvg7Y+qfd5J4Y9OhrJ8Nm4FxYJNNdORaXCsJtTS4yVnAyQoG5sfxfw/dPJrUslmGt6rcymQxM0MMStahMALk4cfM65c9cBSGA7mqmi2k9t/ZZuDJJKbSdpHNjHEdzyRvhtv3G5PyjIbBJOVqDpurP5d+zGWtzeLdahZQLNK895PiV7+N/IURKQQMEqNzKAm1tucnggG4Fkhs4INNuJL145oEuHa7BcIAuWOQRkrhiAFzkkEE8rNDJPczWtnG1rG7yfapvsqESsY1AIJPP3hztbPllTjHKvpv2CWKbTUSHdJCk8cVtH+8jUFOSNpGAynOTgJgL1BNRNxf9fmWNGa3fR7U2ty9zDswJZJxMzHvlwSCc5BwccVerBZ3tLf8Atqyguvsxt/n02OyUSsxk3FsHa24bnyM4ycgE9d1WDDIz1I5BHQ471SM6i1uY8cjjSIWLyFjLbDJu1J5Mefn6HryP4snH3hRdmTS5f7T82aS2WBY5Y5LpVjX94Pn+YYzhmydw4UDB4wio/wDY8K+XLuE1qcfY1B4MWfk6DHOT/Bg4+6K1zGkkHlSoroy7WVlGCCOQRSsU5JGUrPo96EZ5p7O7mmmlubm7XFqQoYIobB2fK3Aztwe3TQvX2QKQSMyxDiQJ1dR1P8u/TvWb5bWty2m3KTXlvqMs7Ky2ieVbqVBKSEcHJL4JB3ZwfeLfcQRnTpY7mT7PLZiOdbJSkillzgA44KMScLsBBAOBkuNxu0/6epcQv9s0+LfJzauxBuVJJDRckdXPJ+YcDJH8QrTrN2sNcsxsk2LZyjd9nXaCWj439VPH3RwcZ/hFaVNGc+gUUUUyAooooAxfFVxZxeG9RS7ltVQ25ytwzbCCQBkJ82MkD5eeah0G7tYLi/sZJrKO4fUbjy4oncM/SQ5D9Ww4J25Xniiis2/eO2FNOhf1/CxmamsEGp2+nzx2CwxajYyWKN9o3KdpX+Ebd2I2287ePm973i3UdMg0W+0y4mtUkNqJPKuBJ5ewuEBYx8gbiBxz+FFFS3ZM1p01OdNN/wBaFfVtU0qHxHplx9qsSkEl2853SvIjxwjdtCZXIX7wbt05q7YJdS6nc6vYxWEtpfi2KS+bKsjQhCSWBBG4bhtAAyCcmiinF3k0RWgqdKMl1VvxZfM2s+VkWFh5nl5x9tfG/djGfK6bec468Y/irN1CeUCeHXbrTItPc3O6GN5PNltxHx0I+YKXLYB42455ooqpaIwo2lO1rE1rZjUYrcCO2Giw/Zp7BYzKrkKu4bh8oAB2EA56HIz0u3uxL+1kbygQpQFt2fmki4GOOuOvfHbNFFPpcTu6nKFpsTU75V8oNJKXIXduOI4hk54zjHTjGO+a0KKKaM59PQxLqNNBMuoQpDFYJEfPQLISp8zcWAXIxh5CcLnOMnHQaFNKkkvtOFoli0k9xqP32Zn24LIFyM7kwRg98c5yUVL3No6xTfV2/IrxyW0vh21kQ2phaexKn97s5aErj+LPTbnj7u7+KujoopxZNZW+9/oQXdpBfWk1rcxiSGVGjdSSMhgQeRyOCRxWOtvbXU0Ol6mIHksJbea0IMoJYISGy3U5ST+JuB83uUUMVJuz121JnaL/AITOBSYPOGnyEA7/ADdvmJnH8G3IHX5s4xxmtmiihdRVFZR9AoooqjI//9k='
|
||||||
|
|
||||||
response = requests.get('https://ocr.holey.cc/ncku?base64_str=%s' % image) #.split('base64,')[1])
|
response = requests.get('https://ocr.holey.cc/ncku?base64_str=%s' % image) # .split('base64,')[1])
|
||||||
print(response.content)
|
print(response.content)
|
|
@ -1,8 +1,10 @@
|
||||||
from curl_cffi import requests
|
|
||||||
from json import loads
|
from json import loads
|
||||||
|
from queue import Queue, Empty
|
||||||
from re import findall
|
from re import findall
|
||||||
from threading import Thread
|
from threading import Thread
|
||||||
from queue import Queue, Empty
|
|
||||||
|
from curl_cffi import requests
|
||||||
|
|
||||||
|
|
||||||
class Completion:
|
class Completion:
|
||||||
# experimental
|
# experimental
|
||||||
|
@ -16,15 +18,16 @@ class Completion:
|
||||||
|
|
||||||
def request():
|
def request():
|
||||||
headers = {
|
headers = {
|
||||||
'authority' : 'chatbot.theb.ai',
|
'authority': 'chatbot.theb.ai',
|
||||||
'content-type': 'application/json',
|
'content-type': 'application/json',
|
||||||
'origin' : 'https://chatbot.theb.ai',
|
'origin': 'https://chatbot.theb.ai',
|
||||||
'user-agent' : 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36',
|
'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36',
|
||||||
}
|
}
|
||||||
|
|
||||||
requests.post('https://chatbot.theb.ai/api/chat-process', headers=headers, content_callback=Completion.handle_stream_response,
|
requests.post('https://chatbot.theb.ai/api/chat-process', headers=headers,
|
||||||
json = {
|
content_callback=Completion.handle_stream_response,
|
||||||
'prompt' : 'hello world',
|
json={
|
||||||
|
'prompt': 'hello world',
|
||||||
'options': {}
|
'options': {}
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
@ -48,10 +51,12 @@ class Completion:
|
||||||
def handle_stream_response(response):
|
def handle_stream_response(response):
|
||||||
Completion.message_queue.put(response.decode())
|
Completion.message_queue.put(response.decode())
|
||||||
|
|
||||||
|
|
||||||
def start():
|
def start():
|
||||||
for message in Completion.create():
|
for message in Completion.create():
|
||||||
yield message['delta']
|
yield message['delta']
|
||||||
|
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
for message in start():
|
for message in start():
|
||||||
print(message)
|
print(message)
|
||||||
|
|
|
@ -1,6 +1,5 @@
|
||||||
import requests
|
import requests
|
||||||
|
|
||||||
|
|
||||||
token = requests.get('https://play.vercel.ai/openai.jpeg', headers={
|
token = requests.get('https://play.vercel.ai/openai.jpeg', headers={
|
||||||
'authority': 'play.vercel.ai',
|
'authority': 'play.vercel.ai',
|
||||||
'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3',
|
'accept-language': 'en,fr-FR;q=0.9,fr;q=0.8,es-ES;q=0.7,es;q=0.6,en-US;q=0.5,am;q=0.4,de;q=0.3',
|
||||||
|
@ -15,7 +14,7 @@ headers = {
|
||||||
'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36'
|
'user-agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/112.0.0.0 Safari/537.36'
|
||||||
}
|
}
|
||||||
|
|
||||||
for chunk in requests.post('https://play.vercel.ai/api/generate', headers=headers, stream = True, json = {
|
for chunk in requests.post('https://play.vercel.ai/api/generate', headers=headers, stream=True, json={
|
||||||
'prompt': 'hi',
|
'prompt': 'hi',
|
||||||
'model': 'openai:gpt-3.5-turbo',
|
'model': 'openai:gpt-3.5-turbo',
|
||||||
'temperature': 0.7,
|
'temperature': 0.7,
|
||||||
|
@ -25,5 +24,4 @@ for chunk in requests.post('https://play.vercel.ai/api/generate', headers=header
|
||||||
'frequencyPenalty': 1,
|
'frequencyPenalty': 1,
|
||||||
'presencePenalty': 1,
|
'presencePenalty': 1,
|
||||||
'stopSequences': []}).iter_lines():
|
'stopSequences': []}).iter_lines():
|
||||||
|
|
||||||
print(chunk)
|
print(chunk)
|
|
@ -1,21 +1,25 @@
|
||||||
from requests import Session
|
|
||||||
from names import get_first_name, get_last_name
|
|
||||||
from random import choice
|
from random import choice
|
||||||
from requests import post
|
|
||||||
from time import time
|
from time import time
|
||||||
from colorama import Fore, init; init()
|
|
||||||
|
from colorama import Fore, init;
|
||||||
|
from names import get_first_name, get_last_name
|
||||||
|
from requests import Session
|
||||||
|
from requests import post
|
||||||
|
|
||||||
|
init()
|
||||||
|
|
||||||
|
|
||||||
class logger:
|
class logger:
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def info(string) -> print:
|
def info(string) -> print:
|
||||||
import datetime
|
import datetime
|
||||||
now = datetime.datetime.now()
|
now = datetime.datetime.now()
|
||||||
return print(f"{Fore.CYAN}{now.strftime('%Y-%m-%d %H:%M:%S')} {Fore.BLUE}INFO {Fore.MAGENTA}__main__ -> {Fore.RESET}{string}")
|
return print(
|
||||||
|
f"{Fore.CYAN}{now.strftime('%Y-%m-%d %H:%M:%S')} {Fore.BLUE}INFO {Fore.MAGENTA}__main__ -> {Fore.RESET}{string}")
|
||||||
|
|
||||||
|
|
||||||
class SonicResponse:
|
class SonicResponse:
|
||||||
|
|
||||||
class Completion:
|
class Completion:
|
||||||
|
|
||||||
class Choices:
|
class Choices:
|
||||||
def __init__(self, choice: dict) -> None:
|
def __init__(self, choice: dict) -> None:
|
||||||
self.text = choice['text']
|
self.text = choice['text']
|
||||||
|
@ -40,7 +44,6 @@ class SonicResponse:
|
||||||
return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>'''
|
return f'''<__main__.APIResponse.Usage(\n prompt_tokens = {self.prompt_tokens},\n completion_tokens = {self.completion_tokens},\n total_tokens = {self.total_tokens})object at 0x1337>'''
|
||||||
|
|
||||||
def __init__(self, response_dict: dict) -> None:
|
def __init__(self, response_dict: dict) -> None:
|
||||||
|
|
||||||
self.response_dict = response_dict
|
self.response_dict = response_dict
|
||||||
self.id = response_dict['id']
|
self.id = response_dict['id']
|
||||||
self.object = response_dict['object']
|
self.object = response_dict['object']
|
||||||
|
@ -52,22 +55,23 @@ class SonicResponse:
|
||||||
def json(self) -> dict:
|
def json(self) -> dict:
|
||||||
return self.response_dict
|
return self.response_dict
|
||||||
|
|
||||||
|
|
||||||
class Account:
|
class Account:
|
||||||
session = Session()
|
session = Session()
|
||||||
session.headers = {
|
session.headers = {
|
||||||
"connection" : "keep-alive",
|
"connection": "keep-alive",
|
||||||
"sec-ch-ua" : "\"Not_A Brand\";v=\"99\", \"Google Chrome\";v=\"109\", \"Chromium\";v=\"109\"",
|
"sec-ch-ua": "\"Not_A Brand\";v=\"99\", \"Google Chrome\";v=\"109\", \"Chromium\";v=\"109\"",
|
||||||
"accept" : "application/json, text/plain, */*",
|
"accept": "application/json, text/plain, */*",
|
||||||
"content-type" : "application/json",
|
"content-type": "application/json",
|
||||||
"sec-ch-ua-mobile" : "?0",
|
"sec-ch-ua-mobile": "?0",
|
||||||
"user-agent" : "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/109.0.0.0 Safari/537.36",
|
"user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/109.0.0.0 Safari/537.36",
|
||||||
"sec-ch-ua-platform": "\"Windows\"",
|
"sec-ch-ua-platform": "\"Windows\"",
|
||||||
"sec-fetch-site" : "same-origin",
|
"sec-fetch-site": "same-origin",
|
||||||
"sec-fetch-mode" : "cors",
|
"sec-fetch-mode": "cors",
|
||||||
"sec-fetch-dest" : "empty",
|
"sec-fetch-dest": "empty",
|
||||||
# "accept-encoding" : "gzip, deflate, br",
|
# "accept-encoding" : "gzip, deflate, br",
|
||||||
"accept-language" : "en-GB,en-US;q=0.9,en;q=0.8",
|
"accept-language": "en-GB,en-US;q=0.9,en;q=0.8",
|
||||||
"cookie" : ""
|
"cookie": ""
|
||||||
}
|
}
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
@ -78,10 +82,10 @@ class Account:
|
||||||
hosts = ['gmail.com', 'protonmail.com', 'proton.me', 'outlook.com']
|
hosts = ['gmail.com', 'protonmail.com', 'proton.me', 'outlook.com']
|
||||||
|
|
||||||
return {
|
return {
|
||||||
"email" : f"{f_name.lower()}.{l_name.lower()}@{choice(hosts)}",
|
"email": f"{f_name.lower()}.{l_name.lower()}@{choice(hosts)}",
|
||||||
"password" : password,
|
"password": password,
|
||||||
"confirm_password" : password,
|
"confirm_password": password,
|
||||||
"full_name" : f'{f_name} {l_name}'
|
"full_name": f'{f_name} {l_name}'
|
||||||
}
|
}
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
|
@ -90,13 +94,13 @@ class Account:
|
||||||
try:
|
try:
|
||||||
user = Account.get_user()
|
user = Account.get_user()
|
||||||
start = time()
|
start = time()
|
||||||
response = Account.session.post("https://app.writesonic.com/api/session-login", json = user | {
|
response = Account.session.post("https://app.writesonic.com/api/session-login", json=user | {
|
||||||
"utmParams" : "{}",
|
"utmParams": "{}",
|
||||||
"visitorId" : "0",
|
"visitorId": "0",
|
||||||
"locale" : "en",
|
"locale": "en",
|
||||||
"userAgent" : "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/109.0.0.0 Safari/537.36",
|
"userAgent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/109.0.0.0 Safari/537.36",
|
||||||
"signInWith" : "password",
|
"signInWith": "password",
|
||||||
"request_type" : "signup",
|
"request_type": "signup",
|
||||||
})
|
})
|
||||||
|
|
||||||
if logging:
|
if logging:
|
||||||
|
@ -105,7 +109,8 @@ class Account:
|
||||||
logger.info(f"\x1b[31mtoken\x1b[0m : '{response.json()['token'][:30]}...'")
|
logger.info(f"\x1b[31mtoken\x1b[0m : '{response.json()['token'][:30]}...'")
|
||||||
|
|
||||||
start = time()
|
start = time()
|
||||||
response = Account.session.post("https://api.writesonic.com/v1/business/set-business-active", headers={"authorization": "Bearer " + response.json()['token']})
|
response = Account.session.post("https://api.writesonic.com/v1/business/set-business-active",
|
||||||
|
headers={"authorization": "Bearer " + response.json()['token']})
|
||||||
key = response.json()["business"]["api_key"]
|
key = response.json()["business"]["api_key"]
|
||||||
if logging: logger.info(f"\x1b[31mgot key\x1b[0m : '{key}' ({int(time() - start)}s)")
|
if logging: logger.info(f"\x1b[31mgot key\x1b[0m : '{key}' ({int(time() - start)}s)")
|
||||||
|
|
||||||
|
@ -129,30 +134,30 @@ class Completion:
|
||||||
enable_memory: bool = False,
|
enable_memory: bool = False,
|
||||||
enable_google_results: bool = False,
|
enable_google_results: bool = False,
|
||||||
history_data: list = []) -> SonicResponse:
|
history_data: list = []) -> SonicResponse:
|
||||||
|
response = post('https://api.writesonic.com/v2/business/content/chatsonic?engine=premium',
|
||||||
response = post('https://api.writesonic.com/v2/business/content/chatsonic?engine=premium', headers = {"X-API-KEY": api_key},
|
headers={"X-API-KEY": api_key},
|
||||||
json = {
|
json={
|
||||||
"enable_memory" : enable_memory,
|
"enable_memory": enable_memory,
|
||||||
"enable_google_results" : enable_google_results,
|
"enable_google_results": enable_google_results,
|
||||||
"input_text" : prompt,
|
"input_text": prompt,
|
||||||
"history_data" : history_data}).json()
|
"history_data": history_data}).json()
|
||||||
|
|
||||||
return SonicResponse({
|
return SonicResponse({
|
||||||
'id' : f'cmpl-premium-{int(time())}',
|
'id': f'cmpl-premium-{int(time())}',
|
||||||
'object' : 'text_completion',
|
'object': 'text_completion',
|
||||||
'created': int(time()),
|
'created': int(time()),
|
||||||
'model' : 'premium',
|
'model': 'premium',
|
||||||
|
|
||||||
'choices': [{
|
'choices': [{
|
||||||
'text' : response['message'],
|
'text': response['message'],
|
||||||
'index' : 0,
|
'index': 0,
|
||||||
'logprobs' : None,
|
'logprobs': None,
|
||||||
'finish_reason' : 'stop'
|
'finish_reason': 'stop'
|
||||||
}],
|
}],
|
||||||
|
|
||||||
'usage': {
|
'usage': {
|
||||||
'prompt_chars' : len(prompt),
|
'prompt_chars': len(prompt),
|
||||||
'completion_chars' : len(response['message']),
|
'completion_chars': len(response['message']),
|
||||||
'total_chars' : len(prompt) + len(response['message'])
|
'total_chars': len(prompt) + len(response['message'])
|
||||||
}
|
}
|
||||||
})
|
})
|
Loading…
Reference in a new issue