Compare commits
No commits in common. "d727d6b68e4d51581052eb1841eb086151612686" and "b1013db0b4f96e279e78b933944bf4e1cb885938" have entirely different histories.
d727d6b68e
...
b1013db0b4
|
@ -234,7 +234,7 @@ func handleAssistantConversation(c tele.Context, thread []*tele.Message) error {
|
|||
}
|
||||
|
||||
req := openai.ChatRequest{
|
||||
Model: openai.ModelGpt41,
|
||||
Model: openai.ModelGpt45,
|
||||
Messages: chatReqMsgs,
|
||||
Temperature: lo.ToPtr(0.42),
|
||||
User: assistantHashUserId(lastMsg.Sender.ID),
|
||||
|
|
|
@ -29,7 +29,7 @@ func handleReasonCmd(c tele.Context) error {
|
|||
}
|
||||
|
||||
req := openai.ChatRequest{
|
||||
Model: openai.ModelO4Mini,
|
||||
Model: openai.ModelO3Mini,
|
||||
Messages: []openai.ChatMessage{
|
||||
{
|
||||
Role: openai.ChatRoleUser,
|
||||
|
|
|
@ -26,7 +26,7 @@ var (
|
|||
"SVCB", "TA", "TKEY", "TLSA", "TSIG", "TXT", "URI", "ZONEMD",
|
||||
})
|
||||
|
||||
errDigInvalidArgs = fmt.Errorf("invalid request")
|
||||
digErrInvalidArgs = fmt.Errorf("invalid request")
|
||||
|
||||
digIdnaMapper = idna.New(idna.MapForLookup(), idna.StrictDomainName(false))
|
||||
)
|
||||
|
@ -43,13 +43,13 @@ func NewDigRequest(req string) (*DigRequest, error) {
|
|||
args := strings.Fields(req)
|
||||
nArgs := len(args)
|
||||
if nArgs == 0 || nArgs > 2 {
|
||||
return nil, errDigInvalidArgs
|
||||
return nil, digErrInvalidArgs
|
||||
}
|
||||
|
||||
if nArgs > 1 {
|
||||
typ := strings.ToUpper(args[1])
|
||||
if _, ok := digValidDnsTypes[typ]; !ok {
|
||||
return nil, errDigInvalidArgs
|
||||
return nil, digErrInvalidArgs
|
||||
}
|
||||
ret.Type = typ
|
||||
}
|
||||
|
@ -64,7 +64,7 @@ func NewDigRequest(req string) (*DigRequest, error) {
|
|||
|
||||
name, err := digIdnaMapper.ToASCII(args[0])
|
||||
if err != nil {
|
||||
return nil, errDigInvalidArgs
|
||||
return nil, digErrInvalidArgs
|
||||
}
|
||||
ret.Name = name
|
||||
return ret, nil
|
||||
|
|
|
@ -69,7 +69,11 @@ func getCachedThread(msg *tele.Message) ([]*tele.Message, error) {
|
|||
|
||||
threadR := []*tele.Message{msg}
|
||||
currentMsg := msg
|
||||
for currentMsg.ReplyTo != nil {
|
||||
for {
|
||||
if currentMsg.ReplyTo == nil {
|
||||
break
|
||||
}
|
||||
|
||||
parentMsg, err := getCachedMessage(currentMsg.ReplyTo)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
|
|
|
@ -67,7 +67,7 @@ func (c *Client) ChatCompletionStream(request ChatRequest) (*ChatResponseStream,
|
|||
}
|
||||
|
||||
if resp.StatusCode() != 200 {
|
||||
defer func() { _ = rbody.Close() }()
|
||||
defer rbody.Close()
|
||||
var respBodyStr string
|
||||
if respBody, err := io.ReadAll(rbody); err == nil {
|
||||
respBodyStr = string(respBody)
|
||||
|
@ -83,7 +83,7 @@ func (c *Client) ChatCompletionStream(request ChatRequest) (*ChatResponseStream,
|
|||
}
|
||||
go func() {
|
||||
defer func() {
|
||||
_ = rbody.Close()
|
||||
rbody.Close()
|
||||
close(ret.Stream)
|
||||
close(ret.Done)
|
||||
}()
|
||||
|
|
|
@ -1,8 +1,9 @@
|
|||
package openai
|
||||
|
||||
const (
|
||||
ModelGpt4O = "gpt-4o" // The safe default, balanced model.
|
||||
ModelO1 = "o1" // Expensive reasoning model
|
||||
ModelO4Mini = "o4-mini" // Cheaper yet powerful reasoning model
|
||||
ModelGpt41 = "gpt-4.1" // OpenAI's Flagship model
|
||||
ModelGpt4O = "gpt-4o" // Safe default
|
||||
ModelO1Preview = "o1-preview" // Expensive reasoning model
|
||||
ModelO1Mini = "o1-mini" // Cheaper reasoning model
|
||||
ModelO3Mini = "o3-mini" // Cheaper yet powerful reasoning model
|
||||
ModelGpt45 = "gpt-4.5-preview" // Bleeding edge (and expensive)
|
||||
)
|
||||
|
|
|
@ -11,7 +11,7 @@ func getLocalIP() (net.IP, error) {
|
|||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
defer func() { _ = conn.Close() }()
|
||||
defer conn.Close()
|
||||
|
||||
host, _, err := net.SplitHostPort(conn.LocalAddr().String())
|
||||
if err != nil {
|
||||
|
|
Loading…
Reference in New Issue