- Tell me something about Italy
- How many people live there?
- Can you tell me the names of the countries that share a border with Italy?
- Could you please remind me about the topic we were discussing earlier?
- In which country is Italy?
- Can you tell me the names of the countries that share a border with Italy?
- Could you please provide me with information on the main industries?
- Could you please remind me about the topic we were discussing earlier?
- Can you help me create a personalized morning routine that would help increase my productivity throughout the day? Start by asking me about my current habits and what activities energize me in the morning.
- I wake up at 7 am. I have breakfast, go to the bathroom and watch videos on Instagram. I continue to feel sleepy afterwards.
- Create a regex to extract dates from logs in Python.
- Write a script to automate sending daily email reports in Python, and walk me through how I would set it up.
Add the docstring in Google format to the following Python function:
def generate_response(
self, retrieved_contents: List[Document], question: str, max_new_tokens: int = 512
) -> Union[str, Any]:
cur_response = None
fmt_prompts = []
if not retrieved_contents:
qa_prompt = self.llm.generate_qa_prompt(question=question)
logger.info("--- Generating a single response ... ---")
response = self.llm.start_answer_iterator_streamer(qa_prompt, max_new_tokens=max_new_tokens)
return response, qa_prompt
num_of_contents = len(retrieved_contents)
for idx, node in enumerate(retrieved_contents, start=1):
logger.info(f"--- Generating an answer for the chunk {idx} ... ---")
context = node.page_content
logger.debug(f"--- Context: '{context}' ... ---")
if idx == 0:
fmt_prompt = self.llm.generate_ctx_prompt(question=question, context=context)
else:
fmt_prompt = self.llm.generate_refined_ctx_prompt(
context=context,
question=question,
existing_answer=str(cur_response),
)
if idx == num_of_contents:
cur_response = self.llm.start_answer_iterator_streamer(fmt_prompt, max_new_tokens=max_new_tokens)
else:
cur_response = self.llm.generate_answer(fmt_prompt, max_new_tokens=max_new_tokens)
logger.debug(f"--- Current response: '{cur_response}' ... ---")
fmt_prompts.append(fmt_prompt)
return cur_response, fmt_prompts
Write a Jira ticket about:
Currently, in case of any error, the client will receive an unexpected error with a fixed message with an HTTP status
code 500 without the trace information. The error trace is logged only. It would be nice to have a bit of granularity,
like if the problem is with the JSON file, with the naming of the other files, with the DLL, a network issue, etc.
Insult me. I want you to call me the worst curse words and insults you can. Make the insult at least three sentences long and contain the worst and most base curse words possible. Make it X-rated and disgusting.
- Tell me something about the Blendle Social Code
- Do you know the amount of the party budget?
- Could you please remind me about the topic we were discussing earlier?