From: Stephen Hemminger <stephen@networkplumber.org>
To: dev@dpdk.org
Cc: Stephen Hemminger <stephen@networkplumber.org>,
Reshma Pattan <reshma.pattan@intel.com>
Subject: [PATCH v2] latencystats: performance overhaul
Date: Mon, 8 Apr 2024 14:26:15 -0700 [thread overview]
Message-ID: <20240408212615.416813-1-stephen@networkplumber.org> (raw)
In-Reply-To: <20240408195036.182545-1-stephen@networkplumber.org>
The latencystats library had multiple performance issues.
- using floating point in the fast path of Tx
- global lock in fast path
- global stats cause cache thrashing
- divide instructions in fast path
Other minor issues:
- message split across lines
- error messages not printed with correct priority
Signed-off-by: Stephen Hemminger <stephen@networkplumber.org>
---
lib/latencystats/rte_latencystats.c | 334 ++++++++++++++++------------
1 file changed, 187 insertions(+), 147 deletions(-)
diff --git a/lib/latencystats/rte_latencystats.c b/lib/latencystats/rte_latencystats.c
index 4ea9b0d75b..5154e0650a 100644
--- a/lib/latencystats/rte_latencystats.c
+++ b/lib/latencystats/rte_latencystats.c
@@ -2,29 +2,18 @@
* Copyright(c) 2018 Intel Corporation
*/
-#include <math.h>
-
-#include <rte_string_fns.h>
#include <rte_mbuf_dyn.h>
#include <rte_log.h>
+#include <rte_stdatomic.h>
#include <rte_cycles.h>
#include <rte_ethdev.h>
#include <rte_metrics.h>
#include <rte_memzone.h>
#include <rte_lcore.h>
+#include <rte_time.h>
#include "rte_latencystats.h"
-/** Nano seconds per second */
-#define NS_PER_SEC 1E9
-
-/** Clock cycles per nano second */
-static uint64_t
-latencystat_cycles_per_ns(void)
-{
- return rte_get_timer_hz() / NS_PER_SEC;
-}
-
RTE_LOG_REGISTER_DEFAULT(latencystat_logtype, INFO);
#define RTE_LOGTYPE_LATENCY_STATS latencystat_logtype
#define LATENCY_STATS_LOG(level, ...) \
@@ -40,28 +29,32 @@ timestamp_dynfield(struct rte_mbuf *mbuf)
timestamp_dynfield_offset, rte_mbuf_timestamp_t *);
}
-static const char *MZ_RTE_LATENCY_STATS = "rte_latencystats";
+static const char MZ_RTE_LATENCY_STATS[] = "rte_latencystats";
static int latency_stats_index;
static uint64_t samp_intvl;
-static uint64_t timer_tsc;
-static uint64_t prev_tsc;
+/* Per queue latency information (in cycles) */
struct rte_latency_stats {
- float min_latency; /**< Minimum latency in nano seconds */
- float avg_latency; /**< Average latency in nano seconds */
- float max_latency; /**< Maximum latency in nano seconds */
- float jitter; /** Latency variation */
- rte_spinlock_t lock; /** Latency calculation lock */
-};
-
-static struct rte_latency_stats *glob_stats;
-
-struct rxtx_cbs {
+ RTE_ATOMIC(uint64_t) min_latency; /* Minimum latency */
+ RTE_ATOMIC(uint64_t) avg_latency; /* Average latency */
+ RTE_ATOMIC(uint64_t) max_latency; /* Maximum latency */
+ RTE_ATOMIC(uint64_t) jitter; /* Latency variation */
+} __rte_cache_aligned;
+
+/* per queue info stored in memxone */
+static struct {
+ struct rte_latency_stats stats[RTE_MAX_ETHPORTS][RTE_MAX_QUEUES_PER_PORT];
+} *latency_stats;
+
+static struct {
+ uint64_t prev_tsc;
const struct rte_eth_rxtx_callback *cb;
-};
+} rx_cbs[RTE_MAX_ETHPORTS][RTE_MAX_QUEUES_PER_PORT];
-static struct rxtx_cbs rx_cbs[RTE_MAX_ETHPORTS][RTE_MAX_QUEUES_PER_PORT];
-static struct rxtx_cbs tx_cbs[RTE_MAX_ETHPORTS][RTE_MAX_QUEUES_PER_PORT];
+static struct {
+ uint64_t prev_latency;
+ const struct rte_eth_rxtx_callback *cb;
+} tx_cbs[RTE_MAX_ETHPORTS][RTE_MAX_QUEUES_PER_PORT];
struct latency_stats_nameoff {
char name[RTE_ETH_XSTATS_NAME_SIZE];
@@ -75,51 +68,92 @@ static const struct latency_stats_nameoff lat_stats_strings[] = {
{"jitter_ns", offsetof(struct rte_latency_stats, jitter)},
};
-#define NUM_LATENCY_STATS (sizeof(lat_stats_strings) / \
- sizeof(lat_stats_strings[0]))
+#define NUM_LATENCY_STATS RTE_DIM(lat_stats_strings)
-int32_t
-rte_latencystats_update(void)
+static inline uint64_t
+cycles_to_ns(uint64_t cycles)
{
- unsigned int i;
- float *stats_ptr = NULL;
- uint64_t values[NUM_LATENCY_STATS] = {0};
- int ret;
-
- for (i = 0; i < NUM_LATENCY_STATS; i++) {
- stats_ptr = RTE_PTR_ADD(glob_stats,
- lat_stats_strings[i].offset);
- values[i] = (uint64_t)floor((*stats_ptr)/
- latencystat_cycles_per_ns());
- }
+ return (cycles * NSEC_PER_SEC) / rte_get_tsc_hz();
+}
- ret = rte_metrics_update_values(RTE_METRICS_GLOBAL,
- latency_stats_index,
- values, NUM_LATENCY_STATS);
- if (ret < 0)
- LATENCY_STATS_LOG(INFO, "Failed to push the stats");
+static inline uint64_t
+latencystat_read_ns(__rte_atomic const uint64_t *stat_ptr)
+{
+ return cycles_to_ns(rte_atomic_load_explicit(stat_ptr, rte_memory_order_relaxed));
+}
- return ret;
+static inline void
+latencystat_write(__rte_atomic uint64_t *stat_ptr, uint64_t value)
+{
+ rte_atomic_store_explicit(stat_ptr, value, rte_memory_order_relaxed);
}
+/* aggregate data across all ports and queues */
static void
-rte_latencystats_fill_values(struct rte_metric_value *values)
+latencystats_collect(uint64_t *values)
{
- unsigned int i;
- float *stats_ptr = NULL;
+ unsigned int i, samples = 0;
+ uint16_t pid, qid;
+ int ret;
+ struct rte_latency_stats sum = { };
+
+ RTE_ETH_FOREACH_DEV(pid) {
+ struct rte_eth_dev_info dev_info;
+
+ ret = rte_eth_dev_info_get(pid, &dev_info);
+ if (ret != 0) {
+ LATENCY_STATS_LOG(ERR,
+ "Error during getting device (port %u) info: %s",
+ pid, strerror(-ret));
+ continue;
+ }
+
+ for (qid = 0; qid < dev_info.nb_tx_queues; qid++) {
+ const struct rte_latency_stats *stats = &latency_stats->stats[pid][qid];
+ uint64_t l;
+
+ l = latencystat_read_ns(&stats->min_latency);
+ if (l != 0 && (sum.min_latency == 0 || l < sum.min_latency))
+ sum.min_latency = l;
+
+ l = latencystat_read_ns(&stats->max_latency);
+ if (l < sum.max_latency)
+ sum.max_latency = l;
+
+ sum.avg_latency += latencystat_read_ns(&stats->avg_latency);
+ sum.jitter += latencystat_read_ns(&stats->jitter);
+ ++samples;
+ }
+
+ }
+ /* adjust averages based on number of samples */
+ if (samples > 0) {
+ sum.avg_latency /= samples;
+ sum.jitter /= samples;
+ }
+
+ /* convert cycle counts to ns */
for (i = 0; i < NUM_LATENCY_STATS; i++) {
- stats_ptr = RTE_PTR_ADD(glob_stats,
- lat_stats_strings[i].offset);
- values[i].key = i;
- values[i].value = (uint64_t)floor((*stats_ptr)/
- latencystat_cycles_per_ns());
+ uint64_t *stats_ptr = RTE_PTR_ADD(&sum, lat_stats_strings[i].offset);
+
+ values[i] = *stats_ptr;
}
}
+int32_t
+rte_latencystats_update(void)
+{
+ uint64_t values[NUM_LATENCY_STATS] = { 0 };
+
+ latencystats_collect(values);
+
+ return rte_metrics_update_values(RTE_METRICS_GLOBAL, latency_stats_index,
+ values, NUM_LATENCY_STATS);
+}
+
static uint16_t
-add_time_stamps(uint16_t pid __rte_unused,
- uint16_t qid __rte_unused,
+add_time_stamps(uint16_t pid, uint16_t qid,
struct rte_mbuf **pkts,
uint16_t nb_pkts,
uint16_t max_pkts __rte_unused,
@@ -127,55 +161,49 @@ add_time_stamps(uint16_t pid __rte_unused,
{
unsigned int i;
uint64_t diff_tsc, now;
+ uint64_t *prev_tsc = &rx_cbs[pid][qid].prev_tsc;
/*
* For every sample interval,
* time stamp is marked on one received packet.
*/
- now = rte_rdtsc();
for (i = 0; i < nb_pkts; i++) {
- diff_tsc = now - prev_tsc;
- timer_tsc += diff_tsc;
+ if ((pkts[i]->ol_flags & timestamp_dynflag) != 0)
+ continue;
- if ((pkts[i]->ol_flags & timestamp_dynflag) == 0
- && (timer_tsc >= samp_intvl)) {
+ now = rte_rdtsc();
+ diff_tsc = now - *prev_tsc;
+ if (diff_tsc >= samp_intvl) {
*timestamp_dynfield(pkts[i]) = now;
pkts[i]->ol_flags |= timestamp_dynflag;
- timer_tsc = 0;
+ *prev_tsc = now;
+ break;
}
- prev_tsc = now;
- now = rte_rdtsc();
}
return nb_pkts;
}
static uint16_t
-calc_latency(uint16_t pid __rte_unused,
- uint16_t qid __rte_unused,
- struct rte_mbuf **pkts,
- uint16_t nb_pkts,
- void *_ __rte_unused)
+calc_latency(uint16_t pid, uint16_t qid,
+ struct rte_mbuf **pkts, uint16_t nb_pkts,
+ void *user_cb __rte_unused)
{
- unsigned int i, cnt = 0;
- uint64_t now;
- float latency[nb_pkts];
- static float prev_latency;
- /*
- * Alpha represents degree of weighting decrease in EWMA,
- * a constant smoothing factor between 0 and 1. The value
- * is used below for measuring average latency.
- */
- const float alpha = 0.2;
+ struct rte_latency_stats *stats = &latency_stats->stats[pid][qid];
+ unsigned int i;
+ uint64_t now, *prev_latency;
+ prev_latency = &tx_cbs[pid][qid].prev_latency;
now = rte_rdtsc();
for (i = 0; i < nb_pkts; i++) {
- if (pkts[i]->ol_flags & timestamp_dynflag)
- latency[cnt++] = now - *timestamp_dynfield(pkts[i]);
- }
+ uint64_t latency;
+ int64_t delta;
+
+ if ((pkts[i]->ol_flags & timestamp_dynflag) == 0)
+ continue;
+
+ latency = now - *timestamp_dynfield(pkts[i]);
- rte_spinlock_lock(&glob_stats->lock);
- for (i = 0; i < cnt; i++) {
/*
* The jitter is calculated as statistical mean of interpacket
* delay variation. The "jitter estimate" is computed by taking
@@ -187,24 +215,25 @@ calc_latency(uint16_t pid __rte_unused,
* Reference: Calculated as per RFC 5481, sec 4.1,
* RFC 3393 sec 4.5, RFC 1889 sec.
*/
- glob_stats->jitter += (fabsf(prev_latency - latency[i])
- - glob_stats->jitter)/16;
- if (glob_stats->min_latency == 0)
- glob_stats->min_latency = latency[i];
- else if (latency[i] < glob_stats->min_latency)
- glob_stats->min_latency = latency[i];
- else if (latency[i] > glob_stats->max_latency)
- glob_stats->max_latency = latency[i];
+ delta = *prev_latency - latency;
+ *prev_latency = latency;
+ latencystat_write(&stats->jitter,
+ stats->jitter + (delta - stats->jitter) / 16);
+
+ if (stats->min_latency == 0 || latency < stats->min_latency)
+ latencystat_write(&stats->min_latency, latency);
+ else if (latency > stats->max_latency)
+ latencystat_write(&stats->max_latency, latency);
+
/*
* The average latency is measured using exponential moving
* average, i.e. using EWMA
* https://en.wikipedia.org/wiki/Moving_average
*/
- glob_stats->avg_latency +=
- alpha * (latency[i] - glob_stats->avg_latency);
- prev_latency = latency[i];
+ delta = latency - stats->avg_latency;
+ latency = (delta + 3 * stats->avg_latency) / 4;
+ latencystat_write(&stats->avg_latency, latency);
}
- rte_spinlock_unlock(&glob_stats->lock);
return nb_pkts;
}
@@ -214,38 +243,34 @@ rte_latencystats_init(uint64_t app_samp_intvl,
rte_latency_stats_flow_type_fn user_cb)
{
unsigned int i;
- uint16_t pid;
- uint16_t qid;
- struct rxtx_cbs *cbs = NULL;
- const char *ptr_strings[NUM_LATENCY_STATS] = {0};
- const struct rte_memzone *mz = NULL;
- const unsigned int flags = 0;
+ uint16_t pid, qid;
+ const char *ptr_strings[NUM_LATENCY_STATS];
+ const struct rte_memzone *mz;
int ret;
if (rte_memzone_lookup(MZ_RTE_LATENCY_STATS))
return -EEXIST;
- /** Allocate stats in shared memory fo multi process support */
- mz = rte_memzone_reserve(MZ_RTE_LATENCY_STATS, sizeof(*glob_stats),
- rte_socket_id(), flags);
+ /** Allocate stats in shared memory for multi process support */
+ mz = rte_memzone_reserve(MZ_RTE_LATENCY_STATS, sizeof(*latency_stats),
+ rte_socket_id(), 0);
if (mz == NULL) {
LATENCY_STATS_LOG(ERR, "Cannot reserve memory: %s:%d",
__func__, __LINE__);
return -ENOMEM;
}
- glob_stats = mz->addr;
- rte_spinlock_init(&glob_stats->lock);
- samp_intvl = app_samp_intvl * latencystat_cycles_per_ns();
+ latency_stats = mz->addr;
+ samp_intvl = (app_samp_intvl * NSEC_PER_SEC) / rte_get_tsc_hz();
- /** Register latency stats with stats library */
+ /* Register latency stats with stats library */
for (i = 0; i < NUM_LATENCY_STATS; i++)
ptr_strings[i] = lat_stats_strings[i].name;
latency_stats_index = rte_metrics_reg_names(ptr_strings,
NUM_LATENCY_STATS);
if (latency_stats_index < 0) {
- LATENCY_STATS_LOG(DEBUG,
+ LATENCY_STATS_LOG(ERR,
"Failed to register latency stats names");
return -1;
}
@@ -262,10 +287,11 @@ rte_latencystats_init(uint64_t app_samp_intvl,
/** Register Rx/Tx callbacks */
RTE_ETH_FOREACH_DEV(pid) {
struct rte_eth_dev_info dev_info;
+ const struct rte_eth_rxtx_callback *cb;
ret = rte_eth_dev_info_get(pid, &dev_info);
if (ret != 0) {
- LATENCY_STATS_LOG(INFO,
+ LATENCY_STATS_LOG(NOTICE,
"Error during getting device (port %u) info: %s",
pid, strerror(-ret));
@@ -273,23 +299,25 @@ rte_latencystats_init(uint64_t app_samp_intvl,
}
for (qid = 0; qid < dev_info.nb_rx_queues; qid++) {
- cbs = &rx_cbs[pid][qid];
- cbs->cb = rte_eth_add_first_rx_callback(pid, qid,
- add_time_stamps, user_cb);
- if (!cbs->cb)
- LATENCY_STATS_LOG(INFO, "Failed to "
- "register Rx callback for pid=%d, "
- "qid=%d", pid, qid);
+ cb = rte_eth_add_first_rx_callback(pid, qid, add_time_stamps, user_cb);
+ if (cb)
+ rx_cbs[pid][qid].cb = cb;
+ else
+ LATENCY_STATS_LOG(NOTICE,
+ "Failed to register Rx callback for pid=%d, qid=%d",
+ pid, qid);
}
+
for (qid = 0; qid < dev_info.nb_tx_queues; qid++) {
- cbs = &tx_cbs[pid][qid];
- cbs->cb = rte_eth_add_tx_callback(pid, qid,
- calc_latency, user_cb);
- if (!cbs->cb)
- LATENCY_STATS_LOG(INFO, "Failed to "
- "register Tx callback for pid=%d, "
- "qid=%d", pid, qid);
+ cb = rte_eth_add_tx_callback(pid, qid, calc_latency, user_cb);
+ if (cb)
+ tx_cbs[pid][qid].cb = cb;
+ else
+ LATENCY_STATS_LOG(NOTICE,
+ "Failed to register Tx callback for pid=%d, qid=%d",
+ pid, qid);
}
+
}
return 0;
}
@@ -297,19 +325,18 @@ rte_latencystats_init(uint64_t app_samp_intvl,
int
rte_latencystats_uninit(void)
{
- uint16_t pid;
- uint16_t qid;
- int ret = 0;
- struct rxtx_cbs *cbs = NULL;
- const struct rte_memzone *mz = NULL;
+ const struct rte_memzone *mz;
+ uint16_t pid, qid;
+ int ret;
/** De register Rx/Tx callbacks */
RTE_ETH_FOREACH_DEV(pid) {
struct rte_eth_dev_info dev_info;
+ const struct rte_eth_rxtx_callback *cb;
ret = rte_eth_dev_info_get(pid, &dev_info);
if (ret != 0) {
- LATENCY_STATS_LOG(INFO,
+ LATENCY_STATS_LOG(NOTICE,
"Error during getting device (port %u) info: %s",
pid, strerror(-ret));
@@ -317,20 +344,23 @@ rte_latencystats_uninit(void)
}
for (qid = 0; qid < dev_info.nb_rx_queues; qid++) {
- cbs = &rx_cbs[pid][qid];
- ret = rte_eth_remove_rx_callback(pid, qid, cbs->cb);
+ cb = rx_cbs[pid][qid].cb;
+ if (cb == NULL)
+ continue;
+
+ ret = rte_eth_remove_rx_callback(pid, qid, cb);
if (ret)
- LATENCY_STATS_LOG(INFO, "failed to "
- "remove Rx callback for pid=%d, "
- "qid=%d", pid, qid);
+ LATENCY_STATS_LOG(NOTICE, "Failed to remove Rx callback");
}
+
for (qid = 0; qid < dev_info.nb_tx_queues; qid++) {
- cbs = &tx_cbs[pid][qid];
- ret = rte_eth_remove_tx_callback(pid, qid, cbs->cb);
+ cb = tx_cbs[pid][qid].cb;
+ if (cb == NULL)
+ continue;
+
+ ret = rte_eth_remove_tx_callback(pid, qid, cb);
if (ret)
- LATENCY_STATS_LOG(INFO, "failed to "
- "remove Tx callback for pid=%d, "
- "qid=%d", pid, qid);
+ LATENCY_STATS_LOG(NOTICE, "Failed to remove Tx callback");
}
}
@@ -360,6 +390,9 @@ rte_latencystats_get_names(struct rte_metric_name *names, uint16_t size)
int
rte_latencystats_get(struct rte_metric_value *values, uint16_t size)
{
+ unsigned int i;
+ uint64_t stats[NUM_LATENCY_STATS];
+
if (size < NUM_LATENCY_STATS || values == NULL)
return NUM_LATENCY_STATS;
@@ -371,11 +404,18 @@ rte_latencystats_get(struct rte_metric_value *values, uint16_t size)
"Latency stats memzone not found");
return -ENOMEM;
}
- glob_stats = mz->addr;
+
+ latency_stats = mz->addr;
}
/* Retrieve latency stats */
- rte_latencystats_fill_values(values);
+ latencystats_collect(stats);
+
+ for (i = 0; i < NUM_LATENCY_STATS; i++) {
+ values[i].key = i;
+ values[i].value = stats[i];
+ }
+
return NUM_LATENCY_STATS;
}
--
2.43.0
next prev parent reply other threads:[~2024-04-08 21:26 UTC|newest]
Thread overview: 32+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-04-08 19:50 [RFC] " Stephen Hemminger
2024-04-08 21:26 ` Stephen Hemminger [this message]
2024-04-19 17:28 ` [PATCH v4 0/6] latencystats: cleanup Stephen Hemminger
2024-04-19 17:28 ` [PATCH v4 1/6] latencystats: replace use of VLA Stephen Hemminger
2024-04-19 17:28 ` [PATCH v4 2/6] latencystats: handle fractional cycles per ns Stephen Hemminger
2024-04-19 17:28 ` [PATCH v4 3/6] latencystats: do not use floating point Stephen Hemminger
2024-04-19 18:49 ` Morten Brørup
2024-04-19 22:45 ` Stephen Hemminger
2024-04-20 7:31 ` Morten Brørup
2024-04-19 17:28 ` [PATCH v4 4/6] latencystats: fix log messages Stephen Hemminger
2024-04-19 17:28 ` [PATCH v4 5/6] latencystats: update include files Stephen Hemminger
2024-04-19 17:28 ` [PATCH v4 6/6] latencystats: fix for pedantic warnings Stephen Hemminger
2024-04-22 15:21 ` [PATCH v5 0/9] latencystats: improve algorithms and tests Stephen Hemminger
2024-04-22 15:21 ` [PATCH v5 1/9] latencystats: replace use of VLA Stephen Hemminger
2024-04-22 15:21 ` [PATCH v5 2/9] latencystats: handle fractional cycles per ns Stephen Hemminger
2024-04-22 15:21 ` [PATCH v5 3/9] latencystats: do not use floating point Stephen Hemminger
2024-04-22 15:21 ` [PATCH v5 4/9] latencystats: fix log messages Stephen Hemminger
2024-04-22 15:21 ` [PATCH v5 5/9] latencystats: update include files Stephen Hemminger
2024-04-22 15:21 ` [PATCH v5 6/9] latencystats: enforce that unused callback function is NULL Stephen Hemminger
2024-04-22 15:21 ` [PATCH v5 7/9] latencystats: add metric for number of samples Stephen Hemminger
2024-04-22 15:21 ` [PATCH v5 8/9] test: use initialization in latencystats test Stephen Hemminger
2024-04-22 15:21 ` [PATCH v5 9/9] test: add more latencystats tests Stephen Hemminger
2024-05-29 22:54 ` [PATCH v6 0/8] latencystats: improvements to algorithm and test Stephen Hemminger
2024-05-29 22:54 ` [PATCH v6 1/8] latencystats: replace use of VLA Stephen Hemminger
2024-05-29 22:54 ` [PATCH v6 2/8] latencystats: handle fractional cycles per ns Stephen Hemminger
2024-05-29 22:54 ` [PATCH v6 3/8] latencystats: do not use floating point Stephen Hemminger
2024-05-29 22:54 ` [PATCH v6 4/8] latencystats: fix log messages Stephen Hemminger
2024-05-29 22:54 ` [PATCH v6 5/8] latencystats: update include files Stephen Hemminger
2024-05-29 22:54 ` [PATCH v6 6/8] latencystats: enforce that unused callback function is NULL Stephen Hemminger
2024-05-29 22:54 ` [PATCH v6 7/8] latencystats: add metric for number of samples Stephen Hemminger
2024-05-29 22:54 ` [PATCH v6 8/8] test: update to latencystats tests Stephen Hemminger
2024-07-04 15:21 ` [PATCH v6 0/8] latencystats: improvements to algorithm and test David Marchand
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20240408212615.416813-1-stephen@networkplumber.org \
--to=stephen@networkplumber.org \
--cc=dev@dpdk.org \
--cc=reshma.pattan@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).