Skip to content

Commit

Permalink
Don't pause processing when send_local_response fails
Browse files Browse the repository at this point in the history
For context see the Envoy issue envoyproxy/envoy#28826.
Here is a shorter summary:

1. A wasm plugin calls proxy_send_local_response from both onRequestHeaders and
   onResponseHeaders
2. When proxy_send_local_reply is called from onRequestHeaders it triggers
   a local reply and that reply goes through the filter chain in Envoy
3. The same plugin is called again as part of the filter chain processing
   but this time onResponseHeaders is called
4. onResponseHeaders calls proxy_send_local_response which ultimately does
   not generate a local reply, but it stops filter chain processing.

As a result we end up with a stuck connection on Envoy - no local reply
and processing is stopped.

I think that proxy wasm plugins shouldn't use proxy_send_local_response this
way, so ultimately whoever created such a plugin shot themselves in the foot.
That being said, I think there are a few improvements that could be made here
on Envoy/proxy-wasm side to handle this situation somewhat better:

1. We can avoid stopping processing in such cases to prevent stuck connections
   on Envoy
2. We can return errors from proxy_send_local_response instead of silently
   ignoring them.

Currently Envoy implementation of sendLocalResponse can detect when a second
local response is requested and returns an error in this case without actually
trying to send a local response.

However, even though Envoy reports an error, send_local_response ignores the
result of the host specific sendLocalResponse implementation and stops
processing and returns success to the wasm plugin.

With this change, send_local_response will check the result of the
host-specific implementation of the sendLocalResponse. In cases when
sendLocalResponse fails it will just propagate the error to the caller and
do nothing else (including stopping processing).

I think this behavior makes sense in principle because on the one hand we don't
ignore the failure from sendLocalResponse and on the other hand, when the
failure happens we don't trigger any side-effects expected from the successful
proxy_send_local_response call.

NOTE: Even though I do think that this is a more resonable behavior, it's
still a change from the previous behavior and it might break existing
proxy-wasm plugins. Specifically:

1. C++ plugins that proactively check the result of proxy_send_local_response
   will change behavior (e.g., before proxy_send_local_response failed silently)
2. Rust plugins, due to the way Rust SDK handles errors from
   proxy_send_local_response will crash in runtime in this case.

On the bright side of things, the plugins that are affected by this change
currently just cause stuck connections in Envoy, so we are changing one
undesirable behavior for another, but more explicit.

Signed-off-by: Mikhail Krinkin <[email protected]>
  • Loading branch information
krinkinmu committed Oct 18, 2024
1 parent 3212034 commit 98fe532
Show file tree
Hide file tree
Showing 6 changed files with 156 additions and 2 deletions.
21 changes: 19 additions & 2 deletions src/exports.cc
Original file line number Diff line number Diff line change
Expand Up @@ -153,8 +153,25 @@ Word send_local_response(Word response_code, Word response_code_details_ptr,
return WasmResult::InvalidMemoryAccess;
}
auto additional_headers = PairsUtil::toPairs(additional_response_header_pairs.value());
context->sendLocalResponse(response_code, body.value(), std::move(additional_headers),
grpc_status, details.value());
auto status = context->sendLocalResponse(
response_code, body.value(), std::move(additional_headers), grpc_status, details.value());
// Only stop processing if we actually triggered local response.
//
// For context, Envoy sends local replies through the filter chain,
// so wasm filter can be called to handle a local reply that the
// filter itself triggered.
//
// Normally that is not an issue, unless wasm filter calls
// proxy_send_local_response again (which they probably shouldn't).
// In this case, no new local response will be generated and
// sendLocalResponse will fail.
//
// If at this point we stop processing, we end up in a situation when
// no response was sent, even though we tried twice, and the connection
// is stuck, because processing is stopped.
if (status != WasmResult::Ok) {
return status;
}
context->wasm()->stopNextIteration(true);
return WasmResult::Ok;
}
Expand Down
1 change: 1 addition & 0 deletions test/BUILD
Original file line number Diff line number Diff line change
Expand Up @@ -89,6 +89,7 @@ cc_test(
data = [
"//test/test_data:clock.wasm",
"//test/test_data:env.wasm",
"//test/test_data:local_response.wasm",
"//test/test_data:random.wasm",
],
linkstatic = 1,
Expand Down
34 changes: 34 additions & 0 deletions test/exports_test.cc
Original file line number Diff line number Diff line change
Expand Up @@ -157,5 +157,39 @@ TEST_P(TestVm, RandomTooLarge) {
EXPECT_TRUE(context->isLogged("random_get(66560) failed."));
}

TEST_P(TestVm, SendLocalResponse) {
auto source = readTestWasmFile("local_response.wasm");
ASSERT_FALSE(source.empty());
auto wasm = TestWasm(std::move(vm_));
ASSERT_TRUE(wasm.load(source, false));
ASSERT_TRUE(wasm.initialize());

auto *context = dynamic_cast<TestContext *>(wasm.vm_context());

// We first try the negative case - proxy_send_local_response fails
WasmCallVoid<0> run_fail;
wasm.wasm_vm()->getFunction("run_fail", &run_fail);
ASSERT_TRUE(run_fail != nullptr);
run_fail(context);

// We expect application to log whatever status
// proxy_send_local_response returns.
EXPECT_TRUE(context->isLogged(stringify("proxy_send_local_response returned ",
static_cast<uint32_t>(WasmResult::Unimplemented))));
// When we fail to send local response we don't pause processing.
EXPECT_FALSE(context->wasm()->isNextIterationStopped());

// Then we try the positive case - proxy_send_local_response succeeds
WasmCallVoid<0> run_success;
wasm.wasm_vm()->getFunction("run_success", &run_success);
ASSERT_TRUE(run_success != nullptr);
run_success(context);

EXPECT_TRUE(context->isLogged(
stringify("proxy_send_local_response returned ", static_cast<uint32_t>(WasmResult::Ok))));
// When we succeed to send local response we stop processing.
EXPECT_TRUE(context->wasm()->isNextIterationStopped());
}

} // namespace
} // namespace proxy_wasm
6 changes: 6 additions & 0 deletions test/test_data/BUILD
Original file line number Diff line number Diff line change
Expand Up @@ -80,6 +80,12 @@ wasm_rust_binary(
wasi = True,
)

wasm_rust_binary(
name = "local_response.wasm",
srcs = ["local_response.rs"],
wasi = True,
)

proxy_wasm_cc_binary(
name = "canary_check.wasm",
srcs = ["canary_check.cc"],
Expand Down
62 changes: 62 additions & 0 deletions test/test_data/local_response.rs
Original file line number Diff line number Diff line change
@@ -0,0 +1,62 @@
// Copyright 2021 Google LLC
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#[no_mangle]
pub extern "C" fn proxy_abi_version_0_2_0() {}

#[no_mangle]
pub extern "C" fn proxy_on_memory_allocate(_: usize) -> *mut u8 {
std::ptr::null_mut()
}

fn send_http_response(status_code: u32) -> u32 {
let headers = 0u32.to_le_bytes().to_vec();
unsafe {
proxy_send_local_response(
status_code,
std::ptr::null(),
0,
std::ptr::null(),
0,
headers.as_ptr(),
headers.len(),
-1)
}
}

#[no_mangle]
pub extern "C" fn run_fail() {
println!(
"proxy_send_local_response returned {}",
send_http_response(404));
}

#[no_mangle]
pub extern "C" fn run_success() {
println!(
"proxy_send_local_response returned {}",
send_http_response(200));
}

extern "C" {
fn proxy_send_local_response(
status_code: u32,
status_code_details_data: *const u8,
status_code_details_size: usize,
body_data: *const u8,
body_size: usize,
headers_data: *const u8,
headers_size: usize,
grpc_status: i32,
) -> u32;
}
34 changes: 34 additions & 0 deletions test/utility.h
Original file line number Diff line number Diff line change
Expand Up @@ -45,6 +45,31 @@ namespace proxy_wasm {
std::vector<std::string> getWasmEngines();
std::string readTestWasmFile(const std::string &filename);

namespace internal {

template <typename... Args> struct Stringify {
static void convert(std::ostream &out) {}
};

template <typename... Args> void stringify_impl(std::ostream &out, Args... args) {
Stringify<Args...>::convert(out, std::forward<Args>(args)...);
}

template <typename A, typename... Args> struct Stringify<A, Args...> {
static void convert(std::ostream &out, A arg, Args... args) {
out << arg;
stringify_impl(out, std::forward<Args>(args)...);
}
};

} // namespace internal

template <typename... Args> std::string stringify(Args... args) {
std::ostringstream out(std::ostringstream::ate);
internal::stringify_impl(out, std::forward<Args>(args)...);
return out.str();
}

class TestIntegration : public WasmVmIntegration {
public:
~TestIntegration() override = default;
Expand Down Expand Up @@ -133,6 +158,15 @@ class TestContext : public ContextBase {
.count();
}

WasmResult sendLocalResponse(uint32_t response_code, std::string_view /*unused*/,
Pairs /*unused*/, GrpcStatusCode /*unused*/,
std::string_view /*unused*/) override {
if (response_code >= 200 && response_code < 300) {
return WasmResult::Ok;
}
return WasmResult::Unimplemented;
}

private:
std::string log_;
static std::string global_log_;
Expand Down

0 comments on commit 98fe532

Please sign in to comment.